ABSTRACT SYNTAX TREE: TOK_QUERY TOK_FROM TOK_SUBQUERY TOK_UNION TOK_UNION TOK_UNION TOK_UNION TOK_UNION TOK_UNION TOK_UNION TOK_UNION TOK_UNION TOK_QUERY TOK_FROM TOK_SUBQUERY TOK_QUERY TOK_FROM TOK_TABREF TOK_TABNAME raw_recreq_dt0 TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECT TOK_SELEXPR TOK_TABLE_OR_COL customer TOK_SELEXPR TOK_TABLE_OR_COL session_id TOK_SELEXPR TOK_FUNCTIONDI count TOK_TABLE_OR_COL gid count_gid_per_session TOK_SELEXPR TOK_FUNCTION count 1 part_pv TOK_WHERE = TOK_TABLE_OR_COL l_date '2015-07-28' TOK_GROUPBY TOK_TABLE_OR_COL customer TOK_TABLE_OR_COL session_id TOK_HAVING <= TOK_TABLE_OR_COL count_gid_per_session 3 t TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECT TOK_SELEXPR '2015-07-28' state_date TOK_SELEXPR TOK_TABLE_OR_COL customer cid TOK_SELEXPR TOK_FUNCTIONDI count TOK_TABLE_OR_COL session_id rec_req_session TOK_SELEXPR TOK_FUNCTIONDI count TOK_TABLE_OR_COL session_id rec_show_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 visits TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_influence_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT TOK_FUNCTION sum TOK_TABLE_OR_COL part_pv rec_req_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT TOK_FUNCTION sum TOK_TABLE_OR_COL part_pv rec_show_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_product_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_product_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_add_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind_po TOK_GROUPBY TOK_TABLE_OR_COL customer TOK_QUERY TOK_FROM TOK_SUBQUERY TOK_QUERY TOK_FROM TOK_SUBQUERY TOK_UNION TOK_QUERY TOK_FROM TOK_SUBQUERY TOK_QUERY TOK_FROM TOK_TABREF TOK_TABNAME raw_event_dt0 TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECT TOK_SELEXPR TOK_TABLE_OR_COL customer TOK_SELEXPR TOK_TABLE_OR_COL gid part_gid TOK_SELEXPR TOK_TABLE_OR_COL session_id part_session_id TOK_WHERE AND = TOK_TABLE_OR_COL l_date '2015-07-28' or = TOK_TABLE_OR_COL version '2' <> TOK_TABLE_OR_COL event_type_id 'PageView' t1 TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECT TOK_SELEXPR TOK_TABLE_OR_COL customer TOK_SELEXPR TOK_TABLE_OR_COL part_session_id TOK_SELEXPR TOK_FUNCTIONDI count TOK_TABLE_OR_COL part_gid count_gid_per_session TOK_GROUPBY TOK_TABLE_OR_COL customer TOK_TABLE_OR_COL part_session_id TOK_HAVING <= TOK_TABLE_OR_COL count_gid_per_session 3 TOK_QUERY TOK_FROM TOK_SUBQUERY TOK_QUERY TOK_FROM TOK_TABREF TOK_TABNAME raw_recreq_dt0 TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECT TOK_SELEXPR TOK_TABLE_OR_COL customer TOK_SELEXPR TOK_TABLE_OR_COL gid part_gid TOK_SELEXPR TOK_TABLE_OR_COL session_id part_session_id TOK_WHERE = TOK_TABLE_OR_COL l_date '2015-07-28' t2 TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECT TOK_SELEXPR TOK_TABLE_OR_COL customer TOK_SELEXPR TOK_TABLE_OR_COL part_session_id TOK_SELEXPR TOK_FUNCTIONDI count TOK_TABLE_OR_COL part_gid count_gid_per_session TOK_GROUPBY TOK_TABLE_OR_COL customer TOK_TABLE_OR_COL part_session_id TOK_HAVING <= TOK_TABLE_OR_COL count_gid_per_session 3 t TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECTDI TOK_SELEXPR TOK_TABLE_OR_COL customer TOK_SELEXPR TOK_TABLE_OR_COL part_session_id t TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECT TOK_SELEXPR '2015-07-28' state_date TOK_SELEXPR TOK_TABLE_OR_COL customer cid TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_session TOK_SELEXPR TOK_FUNCTIONDI count TOK_TABLE_OR_COL part_session_id visits TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_influence_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_product_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_product_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_add_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind_po TOK_GROUPBY TOK_TABLE_OR_COL customer TOK_QUERY TOK_FROM TOK_TABREF TOK_TABNAME raw_kafka_event_order_dt0 TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECT TOK_SELEXPR '2015-07-28' state_date TOK_SELEXPR TOK_TABLE_OR_COL customer cid TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 visits TOK_SELEXPR TOK_FUNCTIONDI count TOK_TABLE_OR_COL session_id purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_influence_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_product_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_product_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_add_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind_po TOK_WHERE = TOK_TABLE_OR_COL l_date '2015-07-28' TOK_GROUPBY TOK_TABLE_OR_COL customer TOK_QUERY TOK_FROM TOK_TABREF TOK_TABNAME raw_kafka_event_feedback_dt0 TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECT TOK_SELEXPR '2015-07-28' state_date TOK_SELEXPR TOK_TABLE_OR_COL customer cid TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 visits TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 purchase_session TOK_SELEXPR TOK_FUNCTIONDI count TOK_TABLE_OR_COL session_id rec_click_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_influence_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_pv TOK_SELEXPR TOK_FUNCTION count 1 rec_click_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_product_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_product_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_add_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind_po TOK_WHERE = TOK_TABLE_OR_COL l_date '2015-07-28' TOK_GROUPBY TOK_TABLE_OR_COL customer TOK_QUERY TOK_FROM TOK_TABREF TOK_TABNAME bi_order_detail_rec_dt0 TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECT TOK_SELEXPR '2015-07-28' state_date TOK_SELEXPR TOK_TABLE_OR_COL customer cid TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 visits TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_session TOK_SELEXPR TOK_FUNCTIONDI count TOK_FUNCTION WHEN >= TOK_TABLE_OR_COL rec_type 9 TOK_TABLE_OR_COL session_id TOK_NULL rec_purchase_session TOK_SELEXPR TOK_FUNCTIONDI count TOK_FUNCTION WHEN >= TOK_TABLE_OR_COL rec_type 3 TOK_TABLE_OR_COL session_id TOK_NULL rec_influence_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_influence_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_product_kind TOK_SELEXPR TOK_FUNCTIONDI count TOK_FUNCTION CASE TOK_TABLE_OR_COL rec_type 9 TOK_TABLE_OR_COL item_id TOK_NULL rec_purchase_product_kind TOK_SELEXPR TOK_FUNCTION round TOK_FUNCTION sum TOK_FUNCTION CASE TOK_TABLE_OR_COL rec_type 9 * TOK_TABLE_OR_COL price TOK_TABLE_OR_COL quantity TOK_NULL 2 rec_purchase_product_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT TOK_FUNCTION sum TOK_FUNCTION CASE TOK_TABLE_OR_COL rec_type 9 TOK_TABLE_OR_COL quantity TOK_NULL rec_purchase_product_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_add_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind_po TOK_WHERE = TOK_TABLE_OR_COL l_date '2015-07-28' TOK_GROUPBY TOK_TABLE_OR_COL customer TOK_QUERY TOK_FROM TOK_TABREF TOK_TABNAME bi_order_detail_rec_dt0 items_rec TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECT TOK_SELEXPR '2015-07-28' state_date TOK_SELEXPR TOK_TABLE_OR_COL customer cid TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 visits TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_purchase_session TOK_SELEXPR TOK_FUNCTIONDI count TOK_TABLE_OR_COL item_id rec_purchase_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT TOK_FUNCTION sum TOK_TABLE_OR_COL quantity rec_purchase_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_influence_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_product_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_product_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_add_kind TOK_SELEXPR TOK_FUNCTIONDI count TOK_TABLE_OR_COL order_id TOK_TABLE_OR_COL item_id rec_purchase_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind_po TOK_SELEXPR TOK_FUNCTIONDI count TOK_FUNCTION CASE TOK_TABLE_OR_COL rec_type 9 TOK_TABLE_OR_COL item_id TOK_NULL rec_purchase_product_kind_po TOK_WHERE AND >= TOK_TABLE_OR_COL rec_type 6 = TOK_TABLE_OR_COL l_date '2015-07-28' TOK_GROUPBY TOK_TABLE_OR_COL customer TOK_QUERY TOK_FROM TOK_TABREF TOK_TABNAME bi_order_detail_rec_dt0 items_rec TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECT TOK_SELEXPR '2015-07-28' state_date TOK_SELEXPR TOK_TABLE_OR_COL customer cid TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 visits TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_products TOK_SELEXPR TOK_FUNCTIONDI count TOK_TABLE_OR_COL item_id rec_influence_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_influence_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT TOK_FUNCTION sum TOK_TABLE_OR_COL quantity rec_influence_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_product_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_product_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_add_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind_po TOK_SELEXPR TOK_FUNCTIONDI count TOK_TABLE_OR_COL order_id TOK_TABLE_OR_COL item_id rec_influence_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind_po TOK_WHERE AND = TOK_TABLE_OR_COL rec_type 3 = TOK_TABLE_OR_COL l_date '2015-07-28' TOK_GROUPBY TOK_TABLE_OR_COL customer TOK_QUERY TOK_FROM TOK_TABREF TOK_TABNAME bi_order_rec_dt0 TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECT TOK_SELEXPR '2015-07-28' state_date TOK_SELEXPR TOK_TABLE_OR_COL customer cid TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 visits TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind TOK_SELEXPR TOK_FUNCTION sum TOK_FUNCTION CASE TOK_TABLE_OR_COL rec_type 4 TOK_TABLE_OR_COL total_price TOK_NULL rec_purchase_order_sales TOK_SELEXPR TOK_FUNCTION count TOK_FUNCTION CASE TOK_TABLE_OR_COL rec_type 4 1 TOK_NULL rec_purchase_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind TOK_SELEXPR TOK_FUNCTION sum TOK_FUNCTION WHEN >= TOK_TABLE_OR_COL rec_type 2 TOK_TABLE_OR_COL total_price TOK_NULL rec_influence_order_sales TOK_SELEXPR TOK_FUNCTION count TOK_FUNCTION WHEN >= TOK_TABLE_OR_COL rec_type 2 1 TOK_NULL rec_influence_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_product_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_product_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_add_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind_po TOK_WHERE = TOK_TABLE_OR_COL l_date '2015-07-28' TOK_GROUPBY TOK_TABLE_OR_COL customer TOK_QUERY TOK_FROM TOK_LEFTOUTERJOIN TOK_SUBQUERY TOK_QUERY TOK_FROM TOK_TABREF TOK_TABNAME bi_order_detail_rec_dt0 TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECTDI TOK_SELEXPR TOK_TABLE_OR_COL customer TOK_SELEXPR TOK_TABLE_OR_COL item_id TOK_WHERE AND = TOK_TABLE_OR_COL l_date '2015-07-28' = TOK_TABLE_OR_COL rec_type 9 created_items_by_rec TOK_SUBQUERY TOK_QUERY TOK_FROM TOK_TABREF TOK_TABNAME bi_order_detail_rec_dt0 TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECTDI TOK_SELEXPR TOK_TABLE_OR_COL customer TOK_SELEXPR TOK_TABLE_OR_COL item_id TOK_WHERE AND = TOK_TABLE_OR_COL l_date '2015-07-28' <> TOK_TABLE_OR_COL rec_type 9 no_relation_with_rec AND = . TOK_TABLE_OR_COL created_items_by_rec customer . TOK_TABLE_OR_COL no_relation_with_rec customer = . TOK_TABLE_OR_COL created_items_by_rec item_id . TOK_TABLE_OR_COL no_relation_with_rec item_id TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECT TOK_SELEXPR '2015-07-28' state_date TOK_SELEXPR . TOK_TABLE_OR_COL created_items_by_rec customer cid TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 visits TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_influence_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_product_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_product_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_amount TOK_SELEXPR TOK_FUNCTIONDI count . TOK_TABLE_OR_COL created_items_by_rec item_id rec_purchase_add_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind_po TOK_WHERE AND TOK_FUNCTION TOK_ISNULL . TOK_TABLE_OR_COL no_relation_with_rec customer TOK_FUNCTION TOK_ISNULL . TOK_TABLE_OR_COL no_relation_with_rec item_id TOK_GROUPBY . TOK_TABLE_OR_COL created_items_by_rec customer TOK_QUERY TOK_FROM TOK_TABREF TOK_TABNAME raw_recresult_dt1 TOK_INSERT TOK_DESTINATION TOK_DIR TOK_TMP_FILE TOK_SELECT TOK_SELEXPR '2015-07-28' state_date TOK_SELEXPR TOK_TABLE_OR_COL customer cid TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 visits TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_purchase_session TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_influence_order_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_products TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_req_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_show_pv TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_click_pv TOK_SELEXPR TOK_FUNCTION count TOK_TABLE_OR_COL rec_item_id rec_show_products TOK_SELEXPR TOK_FUNCTIONDI count TOK_TABLE_OR_COL rec_item_id rec_show_product_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind TOK_SELEXPR TOK_FUNCTION TOK_DOUBLE 0 rec_purchase_product_sales TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_amount TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_add_kind TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind_po TOK_SELEXPR TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind_po TOK_WHERE = TOK_TABLE_OR_COL l_date '2015-07-28' TOK_GROUPBY TOK_TABLE_OR_COL customer t TOK_INSERT TOK_DESTINATION TOK_TAB TOK_TABNAME bi_fact_client_rec_dt0 TOK_PARTSPEC TOK_PARTVAL l_date '2015-07-28' TOK_SELECT TOK_SELEXPR '2015-07-28' TOK_SELEXPR TOK_TABLE_OR_COL cid TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_req_session TOK_FUNCTION TOK_BIGINT 0 rec_req_session TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_show_session TOK_FUNCTION TOK_BIGINT 0 rec_show_session TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL visits TOK_FUNCTION TOK_BIGINT 0 visits TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL purchase_session TOK_FUNCTION TOK_BIGINT 0 purchase_session TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_click_session TOK_FUNCTION TOK_BIGINT 0 rec_click_session TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_purchase_session TOK_FUNCTION TOK_BIGINT 0 rec_purchase_session TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_influence_purchase_session TOK_FUNCTION TOK_BIGINT 0 rec_influence_purchase_session TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_purchase_order_kind TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind TOK_SELEXPR TOK_FUNCTION round TOK_FUNCTION COALESCE TOK_FUNCTION max TOK_TABLE_OR_COL rec_purchase_order_sales TOK_FUNCTION TOK_DOUBLE 0 2 rec_purchase_order_sales TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_purchase_order_amount TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_amount TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_purchase_order_products TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_products TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_influence_order_kind TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind TOK_SELEXPR TOK_FUNCTION round TOK_FUNCTION COALESCE TOK_FUNCTION max TOK_TABLE_OR_COL rec_influence_order_sales TOK_FUNCTION TOK_DOUBLE 0 2 rec_influence_order_sales TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_influence_order_amount TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_amount TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_influence_order_products TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_products TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_req_pv TOK_FUNCTION TOK_BIGINT 0 rec_req_pv TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_show_pv TOK_FUNCTION TOK_BIGINT 0 rec_show_pv TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_click_pv TOK_FUNCTION TOK_BIGINT 0 rec_click_pv TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_show_products TOK_FUNCTION TOK_BIGINT 0 rec_show_products TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_show_product_kind TOK_FUNCTION TOK_BIGINT 0 rec_show_product_kind TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_purchase_product_kind TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind TOK_SELEXPR TOK_FUNCTION round TOK_FUNCTION COALESCE TOK_FUNCTION max TOK_TABLE_OR_COL rec_purchase_product_sales TOK_FUNCTION TOK_DOUBLE 0 2 rec_purchase_product_sales TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_purchase_product_amount TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_amount TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_purchase_add_kind TOK_FUNCTION TOK_BIGINT 0 rec_purchase_add_kind TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_purchase_order_kind_po TOK_FUNCTION TOK_BIGINT 0 rec_purchase_order_kind_po TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_influence_order_kind_po TOK_FUNCTION TOK_BIGINT 0 rec_influence_order_kind_po TOK_SELEXPR TOK_FUNCTION COALESCE TOK_FUNCTION TOK_BIGINT TOK_FUNCTION max TOK_TABLE_OR_COL rec_purchase_product_kind_po TOK_FUNCTION TOK_BIGINT 0 rec_purchase_product_kind_po TOK_WHERE TOK_FUNCTION TOK_ISNOTNULL TOK_TABLE_OR_COL cid TOK_GROUPBY TOK_TABLE_OR_COL cid STAGE DEPENDENCIES: Stage-1 is a root stage Stage-2 depends on stages: Stage-1, Stage-4, Stage-5, Stage-8, Stage-9, Stage-10, Stage-11, Stage-12, Stage-14, Stage-16 Stage-0 depends on stages: Stage-2 Stage-3 depends on stages: Stage-0 Stage-4 is a root stage Stage-5 is a root stage Stage-6 is a root stage Stage-19 depends on stages: Stage-6, Stage-13 , consists of Stage-20, Stage-7 Stage-20 has a backup stage: Stage-7 Stage-18 depends on stages: Stage-20 Stage-8 depends on stages: Stage-7, Stage-18 Stage-7 Stage-9 is a root stage Stage-10 is a root stage Stage-11 is a root stage Stage-12 is a root stage Stage-13 is a root stage Stage-14 is a root stage Stage-15 is a root stage Stage-16 depends on stages: Stage-15, Stage-17 Stage-17 is a root stage STAGE PLANS: Stage: Stage-1 Map Reduce Map Operator Tree: TableScan alias: bi_order_detail_rec_dt0 Statistics: Num rows: 57058 Data size: 18258712 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: customer is not null (type: boolean) Statistics: Num rows: 28529 Data size: 9129356 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: customer (type: string), rec_type (type: int), session_id (type: string), item_id (type: string), price (type: double), quantity (type: double) outputColumnNames: customer, rec_type, session_id, item_id, price, quantity Statistics: Num rows: 28529 Data size: 9129356 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT CASE WHEN ((rec_type >= 9)) THEN (session_id) ELSE (null) END), count(DISTINCT CASE WHEN ((rec_type >= 3)) THEN (session_id) ELSE (null) END), count(DISTINCT CASE (rec_type) WHEN (9) THEN (item_id) ELSE (null) END), sum(CASE (rec_type) WHEN (9) THEN ((price * quantity)) ELSE (null) END), sum(CASE (rec_type) WHEN (9) THEN (quantity) ELSE (null) END) keys: customer (type: string), CASE WHEN ((rec_type >= 9)) THEN (session_id) ELSE (null) END (type: string), CASE WHEN ((rec_type >= 3)) THEN (session_id) ELSE (null) END (type: string), CASE (rec_type) WHEN (9) THEN (item_id) ELSE (null) END (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 Statistics: Num rows: 28529 Data size: 9129356 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) sort order: ++++ Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 28529 Data size: 9129356 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col7 (type: double), _col8 (type: double) auto parallelism: false Path -> Alias: hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:bi_order_detail_rec_dt0] Path -> Partition: hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0/l_date=2015-07-28 Partition base file name: l_date=2015-07-28 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: l_date 2015-07-28 properties: COLUMN_STATS_ACCURATE true bucket_count -1 columns customer,creation_time,user_id,gid,session_id,fid,ip_address,ref_page,item_id,price,quantity,order_id,total_price,version,rec_type columns.comments columns.types string:string:string:string:string:string:string:string:string:double:double:string:double:string:int field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0/l_date=2015-07-28 name default.bi_order_detail_rec_dt0 numFiles 2 numRows 0 partition_columns l_date partition_columns.types string rawDataSize 0 serialization.ddl struct bi_order_detail_rec_dt0 { string customer, string creation_time, string user_id, string gid, string session_id, string fid, string ip_address, string ref_page, string item_id, double price, double quantity, string order_id, double total_price, string version, i32 rec_type} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format totalSize 18258712 transient_lastDdlTime 1438118346 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns customer,creation_time,user_id,gid,session_id,fid,ip_address,ref_page,item_id,price,quantity,order_id,total_price,version,rec_type columns.comments columns.types string:string:string:string:string:string:string:string:string:double:double:string:double:string:int field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0 name default.bi_order_detail_rec_dt0 partition_columns l_date partition_columns.types string serialization.ddl struct bi_order_detail_rec_dt0 { string customer, string creation_time, string user_id, string gid, string session_id, string fid, string ip_address, string ref_page, string item_id, double price, double quantity, string order_id, double total_price, string version, i32 rec_type} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bi_order_detail_rec_dt0 name: default.bi_order_detail_rec_dt0 Truncated Path -> Alias: /bi_order_detail_rec_dt0/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:bi_order_detail_rec_dt0] Needs Tagging: false Reduce Operator Tree: Group By Operator aggregations: count(DISTINCT KEY._col1:0._col0), count(DISTINCT KEY._col1:1._col0), count(DISTINCT KEY._col1:2._col0), sum(VALUE._col3), sum(VALUE._col4) keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 Statistics: Num rows: 14264 Data size: 4564517 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), _col1 (type: bigint), _col2 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), _col3 (type: bigint), round(_col4, 2) (type: double), UDFToLong(_col5) (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 14264 Data size: 4564517 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: true GlobalTableId: 0 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10001 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false Stage: Stage-2 Map Reduce Map Operator Tree: TableScan GatherStats: false Union Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col1 (type: string), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: bigint), _col10 (type: double), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: bigint), _col14 (type: double), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: bigint), _col23 (type: double), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint), _col28 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: max(_col2), max(_col3), max(_col4), max(_col5), max(_col6), max(_col7), max(_col8), max(_col9), max(_col10), max(_col11), max(_col12), max(_col13), max(_col14), max(_col15), max(_col16), max(_col17), max(_col18), max(_col19), max(_col20), max(_col21), max(_col22), max(_col23), max(_col24), max(_col25), max(_col26), max(_col27), max(_col28) keys: _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: double), _col10 (type: bigint), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: double), _col14 (type: bigint), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: double), _col23 (type: bigint), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint) auto parallelism: false TableScan GatherStats: false Union Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col1 (type: string), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: bigint), _col10 (type: double), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: bigint), _col14 (type: double), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: bigint), _col23 (type: double), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint), _col28 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: max(_col2), max(_col3), max(_col4), max(_col5), max(_col6), max(_col7), max(_col8), max(_col9), max(_col10), max(_col11), max(_col12), max(_col13), max(_col14), max(_col15), max(_col16), max(_col17), max(_col18), max(_col19), max(_col20), max(_col21), max(_col22), max(_col23), max(_col24), max(_col25), max(_col26), max(_col27), max(_col28) keys: _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: double), _col10 (type: bigint), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: double), _col14 (type: bigint), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: double), _col23 (type: bigint), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint) auto parallelism: false TableScan GatherStats: false Union Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col1 (type: string), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: bigint), _col10 (type: double), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: bigint), _col14 (type: double), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: bigint), _col23 (type: double), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint), _col28 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: max(_col2), max(_col3), max(_col4), max(_col5), max(_col6), max(_col7), max(_col8), max(_col9), max(_col10), max(_col11), max(_col12), max(_col13), max(_col14), max(_col15), max(_col16), max(_col17), max(_col18), max(_col19), max(_col20), max(_col21), max(_col22), max(_col23), max(_col24), max(_col25), max(_col26), max(_col27), max(_col28) keys: _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: double), _col10 (type: bigint), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: double), _col14 (type: bigint), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: double), _col23 (type: bigint), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint) auto parallelism: false TableScan GatherStats: false Union Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col1 (type: string), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: bigint), _col10 (type: double), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: bigint), _col14 (type: double), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: bigint), _col23 (type: double), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint), _col28 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: max(_col2), max(_col3), max(_col4), max(_col5), max(_col6), max(_col7), max(_col8), max(_col9), max(_col10), max(_col11), max(_col12), max(_col13), max(_col14), max(_col15), max(_col16), max(_col17), max(_col18), max(_col19), max(_col20), max(_col21), max(_col22), max(_col23), max(_col24), max(_col25), max(_col26), max(_col27), max(_col28) keys: _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: double), _col10 (type: bigint), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: double), _col14 (type: bigint), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: double), _col23 (type: bigint), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint) auto parallelism: false TableScan GatherStats: false Union Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col1 (type: string), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: bigint), _col10 (type: double), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: bigint), _col14 (type: double), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: bigint), _col23 (type: double), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint), _col28 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: max(_col2), max(_col3), max(_col4), max(_col5), max(_col6), max(_col7), max(_col8), max(_col9), max(_col10), max(_col11), max(_col12), max(_col13), max(_col14), max(_col15), max(_col16), max(_col17), max(_col18), max(_col19), max(_col20), max(_col21), max(_col22), max(_col23), max(_col24), max(_col25), max(_col26), max(_col27), max(_col28) keys: _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: double), _col10 (type: bigint), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: double), _col14 (type: bigint), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: double), _col23 (type: bigint), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint) auto parallelism: false TableScan GatherStats: false Union Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col1 (type: string), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: bigint), _col10 (type: double), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: bigint), _col14 (type: double), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: bigint), _col23 (type: double), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint), _col28 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: max(_col2), max(_col3), max(_col4), max(_col5), max(_col6), max(_col7), max(_col8), max(_col9), max(_col10), max(_col11), max(_col12), max(_col13), max(_col14), max(_col15), max(_col16), max(_col17), max(_col18), max(_col19), max(_col20), max(_col21), max(_col22), max(_col23), max(_col24), max(_col25), max(_col26), max(_col27), max(_col28) keys: _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: double), _col10 (type: bigint), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: double), _col14 (type: bigint), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: double), _col23 (type: bigint), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint) auto parallelism: false TableScan GatherStats: false Union Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col1 (type: string), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: bigint), _col10 (type: double), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: bigint), _col14 (type: double), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: bigint), _col23 (type: double), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint), _col28 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: max(_col2), max(_col3), max(_col4), max(_col5), max(_col6), max(_col7), max(_col8), max(_col9), max(_col10), max(_col11), max(_col12), max(_col13), max(_col14), max(_col15), max(_col16), max(_col17), max(_col18), max(_col19), max(_col20), max(_col21), max(_col22), max(_col23), max(_col24), max(_col25), max(_col26), max(_col27), max(_col28) keys: _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: double), _col10 (type: bigint), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: double), _col14 (type: bigint), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: double), _col23 (type: bigint), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint) auto parallelism: false TableScan GatherStats: false Union Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col1 (type: string), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: bigint), _col10 (type: double), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: bigint), _col14 (type: double), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: bigint), _col23 (type: double), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint), _col28 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: max(_col2), max(_col3), max(_col4), max(_col5), max(_col6), max(_col7), max(_col8), max(_col9), max(_col10), max(_col11), max(_col12), max(_col13), max(_col14), max(_col15), max(_col16), max(_col17), max(_col18), max(_col19), max(_col20), max(_col21), max(_col22), max(_col23), max(_col24), max(_col25), max(_col26), max(_col27), max(_col28) keys: _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: double), _col10 (type: bigint), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: double), _col14 (type: bigint), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: double), _col23 (type: bigint), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint) auto parallelism: false TableScan GatherStats: false Union Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col1 (type: string), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: bigint), _col10 (type: double), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: bigint), _col14 (type: double), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: bigint), _col23 (type: double), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint), _col28 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: max(_col2), max(_col3), max(_col4), max(_col5), max(_col6), max(_col7), max(_col8), max(_col9), max(_col10), max(_col11), max(_col12), max(_col13), max(_col14), max(_col15), max(_col16), max(_col17), max(_col18), max(_col19), max(_col20), max(_col21), max(_col22), max(_col23), max(_col24), max(_col25), max(_col26), max(_col27), max(_col28) keys: _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: double), _col10 (type: bigint), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: double), _col14 (type: bigint), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: double), _col23 (type: bigint), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint) auto parallelism: false TableScan GatherStats: false Union Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col1 (type: string), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: bigint), _col10 (type: double), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: bigint), _col14 (type: double), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: bigint), _col23 (type: double), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint), _col28 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: max(_col2), max(_col3), max(_col4), max(_col5), max(_col6), max(_col7), max(_col8), max(_col9), max(_col10), max(_col11), max(_col12), max(_col13), max(_col14), max(_col15), max(_col16), max(_col17), max(_col18), max(_col19), max(_col20), max(_col21), max(_col22), max(_col23), max(_col24), max(_col25), max(_col26), max(_col27), max(_col28) keys: _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27 Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 933929860 Data size: 188405722027 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: bigint), _col5 (type: bigint), _col6 (type: bigint), _col7 (type: bigint), _col8 (type: bigint), _col9 (type: double), _col10 (type: bigint), _col11 (type: bigint), _col12 (type: bigint), _col13 (type: double), _col14 (type: bigint), _col15 (type: bigint), _col16 (type: bigint), _col17 (type: bigint), _col18 (type: bigint), _col19 (type: bigint), _col20 (type: bigint), _col21 (type: bigint), _col22 (type: double), _col23 (type: bigint), _col24 (type: bigint), _col25 (type: bigint), _col26 (type: bigint), _col27 (type: bigint) auto parallelism: false Path -> Alias: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10001 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10001] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10002 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10002] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10003 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10003] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10006 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10006] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10007 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10007] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10008 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10008] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10009 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10009] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10010 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10010] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10012 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10012] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10014 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10014] Path -> Partition: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10001 Partition base file name: -mr-10001 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10002 Partition base file name: -mr-10002 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10003 Partition base file name: -mr-10003 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10006 Partition base file name: -mr-10006 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10007 Partition base file name: -mr-10007 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10008 Partition base file name: -mr-10008 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10009 Partition base file name: -mr-10009 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10010 Partition base file name: -mr-10010 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10012 Partition base file name: -mr-10012 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10014 Partition base file name: -mr-10014 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe Truncated Path -> Alias: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10001 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10001] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10002 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10002] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10003 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10003] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10006 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10006] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10007 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10007] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10008 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10008] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10009 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10009] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10010 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10010] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10012 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10012] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10014 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10014] Needs Tagging: false Reduce Operator Tree: Group By Operator aggregations: max(VALUE._col0), max(VALUE._col1), max(VALUE._col2), max(VALUE._col3), max(VALUE._col4), max(VALUE._col5), max(VALUE._col6), max(VALUE._col7), max(VALUE._col8), max(VALUE._col9), max(VALUE._col10), max(VALUE._col11), max(VALUE._col12), max(VALUE._col13), max(VALUE._col14), max(VALUE._col15), max(VALUE._col16), max(VALUE._col17), max(VALUE._col18), max(VALUE._col19), max(VALUE._col20), max(VALUE._col21), max(VALUE._col22), max(VALUE._col23), max(VALUE._col24), max(VALUE._col25), max(VALUE._col26) keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27 Statistics: Num rows: 466964930 Data size: 94202861013 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: '2015-07-28' (type: string), _col0 (type: string), COALESCE(_col1,0) (type: bigint), COALESCE(_col2,0) (type: bigint), COALESCE(_col3,0) (type: bigint), COALESCE(_col4,0) (type: bigint), COALESCE(_col5,0) (type: bigint), COALESCE(_col6,0) (type: bigint), COALESCE(_col7,0) (type: bigint), COALESCE(_col8,0) (type: bigint), round(COALESCE(_col9,0.0), 2) (type: double), COALESCE(_col10,0) (type: bigint), COALESCE(_col11,0) (type: bigint), COALESCE(_col12,0) (type: bigint), round(COALESCE(_col13,0.0), 2) (type: double), COALESCE(_col14,0) (type: bigint), COALESCE(_col15,0) (type: bigint), COALESCE(_col16,0) (type: bigint), COALESCE(_col17,0) (type: bigint), COALESCE(_col18,0) (type: bigint), COALESCE(_col19,0) (type: bigint), COALESCE(_col20,0) (type: bigint), COALESCE(_col21,0) (type: bigint), round(COALESCE(_col22,0.0), 2) (type: double), COALESCE(_col23,0) (type: bigint), COALESCE(_col24,0) (type: bigint), COALESCE(_col25,0) (type: bigint), COALESCE(_col26,0) (type: bigint), COALESCE(_col27,0) (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 466964930 Data size: 94202861013 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false GlobalTableId: 1 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-ext-10000 NumFilesPerFileSink: 1 Static Partition Specification: l_date=2015-07-28/ Statistics: Num rows: 466964930 Data size: 94202861013 Basic stats: COMPLETE Column stats: NONE Stats Publishing Key Prefix: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-ext-10000/ table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns state_date,cid,rec_req_session,rec_show_session,visits,purchase_session,rec_click_session,rec_purchase_session,rec_influence_purchase_session,rec_purchase_order_kind,rec_purchase_order_sales,rec_purchase_order_amount,rec_purchase_order_products,rec_influence_order_kind,rec_influence_order_sales,rec_influence_order_amount,rec_influence_order_products,rec_req_pv,rec_show_pv,rec_click_pv,rec_show_products,rec_show_product_kind,rec_purchase_product_kind,rec_purchase_product_sales,rec_purchase_product_amount,rec_purchase_add_kind,rec_purchase_order_kind_po,rec_influence_order_kind_po,rec_purchase_product_kind_po columns.comments columns.types string:string:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:double:bigint:bigint:bigint:double:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:double:bigint:bigint:bigint:bigint:bigint field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/bi_fact_client_rec_dt0 name default.bi_fact_client_rec_dt0 partition_columns l_date partition_columns.types string serialization.ddl struct bi_fact_client_rec_dt0 { string state_date, string cid, i64 rec_req_session, i64 rec_show_session, i64 visits, i64 purchase_session, i64 rec_click_session, i64 rec_purchase_session, i64 rec_influence_purchase_session, i64 rec_purchase_order_kind, double rec_purchase_order_sales, i64 rec_purchase_order_amount, i64 rec_purchase_order_products, i64 rec_influence_order_kind, double rec_influence_order_sales, i64 rec_influence_order_amount, i64 rec_influence_order_products, i64 rec_req_pv, i64 rec_show_pv, i64 rec_click_pv, i64 rec_show_products, i64 rec_show_product_kind, i64 rec_purchase_product_kind, double rec_purchase_product_sales, i64 rec_purchase_product_amount, i64 rec_purchase_add_kind, i64 rec_purchase_order_kind_po, i64 rec_influence_order_kind_po, i64 rec_purchase_product_kind_po} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bi_fact_client_rec_dt0 TotalFiles: 1 GatherStats: true MultiFileSpray: false Stage: Stage-0 Move Operator tables: partition: l_date 2015-07-28 replace: true source: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-ext-10000 table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns state_date,cid,rec_req_session,rec_show_session,visits,purchase_session,rec_click_session,rec_purchase_session,rec_influence_purchase_session,rec_purchase_order_kind,rec_purchase_order_sales,rec_purchase_order_amount,rec_purchase_order_products,rec_influence_order_kind,rec_influence_order_sales,rec_influence_order_amount,rec_influence_order_products,rec_req_pv,rec_show_pv,rec_click_pv,rec_show_products,rec_show_product_kind,rec_purchase_product_kind,rec_purchase_product_sales,rec_purchase_product_amount,rec_purchase_add_kind,rec_purchase_order_kind_po,rec_influence_order_kind_po,rec_purchase_product_kind_po columns.comments columns.types string:string:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:double:bigint:bigint:bigint:double:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:double:bigint:bigint:bigint:bigint:bigint field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/bi_fact_client_rec_dt0 name default.bi_fact_client_rec_dt0 partition_columns l_date partition_columns.types string serialization.ddl struct bi_fact_client_rec_dt0 { string state_date, string cid, i64 rec_req_session, i64 rec_show_session, i64 visits, i64 purchase_session, i64 rec_click_session, i64 rec_purchase_session, i64 rec_influence_purchase_session, i64 rec_purchase_order_kind, double rec_purchase_order_sales, i64 rec_purchase_order_amount, i64 rec_purchase_order_products, i64 rec_influence_order_kind, double rec_influence_order_sales, i64 rec_influence_order_amount, i64 rec_influence_order_products, i64 rec_req_pv, i64 rec_show_pv, i64 rec_click_pv, i64 rec_show_products, i64 rec_show_product_kind, i64 rec_purchase_product_kind, double rec_purchase_product_sales, i64 rec_purchase_product_amount, i64 rec_purchase_add_kind, i64 rec_purchase_order_kind_po, i64 rec_influence_order_kind_po, i64 rec_purchase_product_kind_po} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bi_fact_client_rec_dt0 Stage: Stage-3 Stats-Aggr Operator Stats Aggregation Key Prefix: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-ext-10000/ Stage: Stage-4 Map Reduce Map Operator Tree: TableScan alias: raw_kafka_event_feedback_dt0 Statistics: Num rows: 2780483 Data size: 556096615 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: customer is not null (type: boolean) Statistics: Num rows: 1390242 Data size: 278048407 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: customer (type: string), session_id (type: string) outputColumnNames: customer, session_id Statistics: Num rows: 1390242 Data size: 278048407 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT session_id), count(1) keys: customer (type: string), session_id (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1390242 Data size: 278048407 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 1390242 Data size: 278048407 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col3 (type: bigint) auto parallelism: false Path -> Alias: hdfs://bfdhadoopcool/warehouse/raw_kafka_event_feedback_dt0/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:raw_kafka_event_feedback_dt0] Path -> Partition: hdfs://bfdhadoopcool/warehouse/raw_kafka_event_feedback_dt0/l_date=2015-07-28 Partition base file name: l_date=2015-07-28 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: l_date 2015-07-28 properties: COLUMN_STATS_ACCURATE true bucket_count -1 columns customer,uid,session_id,gid,ip_address,event_type_id,ref_page,creation_time,tma,tmc,tmd,item_id,cat_id,page_t,rid,app,rv,src,b_on,b_pn,attr,bid,extend columns.comments columns.types string:string:string:string:string:string:string:string:string:string:string:string:string:string:string:string:double:string:string:bigint:map:string:map field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/raw_kafka_event_feedback_dt0/l_date=2015-07-28 name default.raw_kafka_event_feedback_dt0 numFiles 3 numRows 0 partition_columns l_date partition_columns.types string rawDataSize 0 serialization.ddl struct raw_kafka_event_feedback_dt0 { string customer, string uid, string session_id, string gid, string ip_address, string event_type_id, string ref_page, string creation_time, string tma, string tmc, string tmd, string item_id, string cat_id, string page_t, string rid, string app, double rv, string src, string b_on, i64 b_pn, map attr, string bid, map extend} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format totalSize 556096615 transient_lastDdlTime 1438108241 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns customer,uid,session_id,gid,ip_address,event_type_id,ref_page,creation_time,tma,tmc,tmd,item_id,cat_id,page_t,rid,app,rv,src,b_on,b_pn,attr,bid,extend columns.comments columns.types string:string:string:string:string:string:string:string:string:string:string:string:string:string:string:string:double:string:string:bigint:map:string:map field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/raw_kafka_event_feedback_dt0 name default.raw_kafka_event_feedback_dt0 partition_columns l_date partition_columns.types string serialization.ddl struct raw_kafka_event_feedback_dt0 { string customer, string uid, string session_id, string gid, string ip_address, string event_type_id, string ref_page, string creation_time, string tma, string tmc, string tmd, string item_id, string cat_id, string page_t, string rid, string app, double rv, string src, string b_on, i64 b_pn, map attr, string bid, map extend} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.raw_kafka_event_feedback_dt0 name: default.raw_kafka_event_feedback_dt0 Truncated Path -> Alias: /raw_kafka_event_feedback_dt0/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:raw_kafka_event_feedback_dt0] Needs Tagging: false Reduce Operator Tree: Group By Operator aggregations: count(DISTINCT KEY._col1:0._col0), count(VALUE._col1) keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 695121 Data size: 139024203 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), _col1 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), _col2 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 695121 Data size: 139024203 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: true GlobalTableId: 0 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10002 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false Stage: Stage-5 Map Reduce Map Operator Tree: TableScan alias: bi_order_rec_dt0 Statistics: Num rows: 73229 Data size: 8201738 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: customer is not null (type: boolean) Statistics: Num rows: 36615 Data size: 4100925 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: customer (type: string), rec_type (type: int), total_price (type: double) outputColumnNames: customer, rec_type, total_price Statistics: Num rows: 36615 Data size: 4100925 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: sum(CASE (rec_type) WHEN (4) THEN (total_price) ELSE (null) END), count(CASE (rec_type) WHEN (4) THEN (1) ELSE (null) END), sum(CASE WHEN ((rec_type >= 2)) THEN (total_price) ELSE (null) END), count(CASE WHEN ((rec_type >= 2)) THEN (1) ELSE (null) END) keys: customer (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4 Statistics: Num rows: 36615 Data size: 4100925 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string) sort order: + Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 36615 Data size: 4100925 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col1 (type: double), _col2 (type: bigint), _col3 (type: double), _col4 (type: bigint) auto parallelism: false Path -> Alias: hdfs://bfdhadoopcool/warehouse/bi_order_rec_dt0/l_date=2015-07-28 [null-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery2:bi_order_rec_dt0] Path -> Partition: hdfs://bfdhadoopcool/warehouse/bi_order_rec_dt0/l_date=2015-07-28 Partition base file name: l_date=2015-07-28 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: l_date 2015-07-28 properties: COLUMN_STATS_ACCURATE true bucket_count -1 columns customer,creation_time,user_id,gid,session_id,fid,ip_address,ref_page,order_id,total_price,version,rec_type columns.comments columns.types string:string:string:string:string:string:string:string:string:double:string:int field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/bi_order_rec_dt0/l_date=2015-07-28 name default.bi_order_rec_dt0 numFiles 1 numRows 0 partition_columns l_date partition_columns.types string rawDataSize 0 serialization.ddl struct bi_order_rec_dt0 { string customer, string creation_time, string user_id, string gid, string session_id, string fid, string ip_address, string ref_page, string order_id, double total_price, string version, i32 rec_type} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format totalSize 8201738 transient_lastDdlTime 1438118552 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns customer,creation_time,user_id,gid,session_id,fid,ip_address,ref_page,order_id,total_price,version,rec_type columns.comments columns.types string:string:string:string:string:string:string:string:string:double:string:int field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/bi_order_rec_dt0 name default.bi_order_rec_dt0 partition_columns l_date partition_columns.types string serialization.ddl struct bi_order_rec_dt0 { string customer, string creation_time, string user_id, string gid, string session_id, string fid, string ip_address, string ref_page, string order_id, double total_price, string version, i32 rec_type} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bi_order_rec_dt0 name: default.bi_order_rec_dt0 Truncated Path -> Alias: /bi_order_rec_dt0/l_date=2015-07-28 [null-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery2:bi_order_rec_dt0] Needs Tagging: false Reduce Operator Tree: Group By Operator aggregations: sum(VALUE._col0), count(VALUE._col1), sum(VALUE._col2), count(VALUE._col3) keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3, _col4 Statistics: Num rows: 18307 Data size: 2050406 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), _col1 (type: double), _col2 (type: bigint), 0 (type: bigint), 0 (type: bigint), _col3 (type: double), _col4 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 18307 Data size: 2050406 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: true GlobalTableId: 0 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10003 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false Stage: Stage-6 Map Reduce Map Operator Tree: TableScan alias: bi_order_detail_rec_dt0 Statistics: Num rows: 89503 Data size: 18258712 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: ((rec_type <> 9) and customer is not null) (type: boolean) Statistics: Num rows: 44752 Data size: 9129458 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: customer (type: string), item_id (type: string) outputColumnNames: customer, item_id Statistics: Num rows: 44752 Data size: 9129458 Basic stats: COMPLETE Column stats: NONE Group By Operator keys: customer (type: string), item_id (type: string) mode: hash outputColumnNames: _col0, _col1 Statistics: Num rows: 44752 Data size: 9129458 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) Statistics: Num rows: 44752 Data size: 9129458 Basic stats: COMPLETE Column stats: NONE tag: -1 auto parallelism: false Path -> Alias: hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0/l_date=2015-07-28 [null-subquery1-subquery2:t-subquery1-subquery2:no_relation_with_rec:bi_order_detail_rec_dt0] Path -> Partition: hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0/l_date=2015-07-28 Partition base file name: l_date=2015-07-28 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: l_date 2015-07-28 properties: COLUMN_STATS_ACCURATE true bucket_count -1 columns customer,creation_time,user_id,gid,session_id,fid,ip_address,ref_page,item_id,price,quantity,order_id,total_price,version,rec_type columns.comments columns.types string:string:string:string:string:string:string:string:string:double:double:string:double:string:int field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0/l_date=2015-07-28 name default.bi_order_detail_rec_dt0 numFiles 2 numRows 0 partition_columns l_date partition_columns.types string rawDataSize 0 serialization.ddl struct bi_order_detail_rec_dt0 { string customer, string creation_time, string user_id, string gid, string session_id, string fid, string ip_address, string ref_page, string item_id, double price, double quantity, string order_id, double total_price, string version, i32 rec_type} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format totalSize 18258712 transient_lastDdlTime 1438118346 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns customer,creation_time,user_id,gid,session_id,fid,ip_address,ref_page,item_id,price,quantity,order_id,total_price,version,rec_type columns.comments columns.types string:string:string:string:string:string:string:string:string:double:double:string:double:string:int field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0 name default.bi_order_detail_rec_dt0 partition_columns l_date partition_columns.types string serialization.ddl struct bi_order_detail_rec_dt0 { string customer, string creation_time, string user_id, string gid, string session_id, string fid, string ip_address, string ref_page, string item_id, double price, double quantity, string order_id, double total_price, string version, i32 rec_type} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bi_order_detail_rec_dt0 name: default.bi_order_detail_rec_dt0 Truncated Path -> Alias: /bi_order_detail_rec_dt0/l_date=2015-07-28 [null-subquery1-subquery2:t-subquery1-subquery2:no_relation_with_rec:bi_order_detail_rec_dt0] Needs Tagging: false Reduce Operator Tree: Group By Operator keys: KEY._col0 (type: string), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 22376 Data size: 4564729 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 22376 Data size: 4564729 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: true GlobalTableId: 0 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10004 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1 columns.types string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false Stage: Stage-19 Conditional Operator Stage: Stage-20 Map Reduce Local Work Alias -> Map Local Tables: null-subquery1-subquery2:t-subquery1-subquery2:$INTNAME Fetch Operator limit: -1 Alias -> Map Local Operator Tree: null-subquery1-subquery2:t-subquery1-subquery2:$INTNAME TableScan GatherStats: false HashTable Sink Operator condition expressions: 0 {_col0} {_col1} 1 keys: 0 _col0 (type: string), _col1 (type: string) 1 _col0 (type: string), _col1 (type: string) Position of Big Table: 0 Stage: Stage-18 Map Reduce Map Operator Tree: TableScan GatherStats: false Map Join Operator condition map: Left Outer Join0 to 1 condition expressions: 0 {_col0} {_col1} 1 {_col0} {_col1} keys: 0 _col0 (type: string), _col1 (type: string) 1 _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Position of Big Table: 0 Filter Operator isSamplingPred: false predicate: (_col2 is null and _col3 is null) (type: boolean) Select Operator expressions: _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1 Group By Operator aggregations: count(DISTINCT _col1) keys: _col0 (type: string), _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2 File Output Operator compressed: true GlobalTableId: 0 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10005 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1,_col2 columns.types string,string,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false Local Work: Map Reduce Local Work Path -> Alias: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10011 [null-subquery1-subquery2:t-subquery1-subquery2:$INTNAME1] Path -> Partition: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10004 Partition base file name: -mr-10004 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1 columns.types string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1 columns.types string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10011 Partition base file name: -mr-10011 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1 columns.types string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1 columns.types string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe Truncated Path -> Alias: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10011 [null-subquery1-subquery2:t-subquery1-subquery2:$INTNAME1] Stage: Stage-8 Map Reduce Map Operator Tree: TableScan GatherStats: false Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 6153 Data size: 1255249 Basic stats: COMPLETE Column stats: NONE tag: -1 auto parallelism: false Path -> Alias: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10005 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10005] Path -> Partition: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10005 Partition base file name: -mr-10005 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1,_col2 columns.types string,string,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1,_col2 columns.types string,string,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe Truncated Path -> Alias: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10005 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10005] Needs Tagging: false Reduce Operator Tree: Group By Operator aggregations: count(DISTINCT KEY._col1:0._col0) keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 3076 Data size: 627522 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), _col1 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 3076 Data size: 627522 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: true GlobalTableId: 0 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10006 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false Stage: Stage-7 Map Reduce Map Operator Tree: TableScan GatherStats: false Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) Statistics: Num rows: 22376 Data size: 4564729 Basic stats: COMPLETE Column stats: NONE tag: 1 auto parallelism: false TableScan GatherStats: false Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) Statistics: Num rows: 11188 Data size: 2282364 Basic stats: COMPLETE Column stats: NONE tag: 0 auto parallelism: false Path -> Alias: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10004 [null-subquery1-subquery2:t-subquery1-subquery2:$INTNAME] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10011 [null-subquery1-subquery2:t-subquery1-subquery2:$INTNAME1] Path -> Partition: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10004 Partition base file name: -mr-10004 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1 columns.types string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1 columns.types string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10011 Partition base file name: -mr-10011 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1 columns.types string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1 columns.types string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe Truncated Path -> Alias: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10004 [null-subquery1-subquery2:t-subquery1-subquery2:$INTNAME] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10011 [null-subquery1-subquery2:t-subquery1-subquery2:$INTNAME1] Needs Tagging: true Reduce Operator Tree: Join Operator condition map: Left Outer Join0 to 1 condition expressions: 0 {KEY.reducesinkkey0} {KEY.reducesinkkey1} 1 {KEY.reducesinkkey0} {KEY.reducesinkkey1} outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 24613 Data size: 5021202 Basic stats: COMPLETE Column stats: NONE Filter Operator isSamplingPred: false predicate: (_col2 is null and _col3 is null) (type: boolean) Statistics: Num rows: 6153 Data size: 1255249 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 6153 Data size: 1255249 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT _col1) keys: _col0 (type: string), _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 6153 Data size: 1255249 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: true GlobalTableId: 0 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10005 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1,_col2 columns.types string,string,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false Stage: Stage-9 Map Reduce Map Operator Tree: TableScan alias: raw_recreq_dt0 Statistics: Num rows: 100267659 Data size: 30080297905 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: customer is not null (type: boolean) Statistics: Num rows: 50133830 Data size: 15040149102 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: customer (type: string), session_id (type: string), gid (type: string) outputColumnNames: customer, session_id, gid Statistics: Num rows: 50133830 Data size: 15040149102 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT gid), count(1) keys: customer (type: string), session_id (type: string), gid (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4 Statistics: Num rows: 50133830 Data size: 15040149102 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 50133830 Data size: 15040149102 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col4 (type: bigint) auto parallelism: false Path -> Alias: hdfs://bfdhadoopcool/warehouse/raw_recreq_dt0/l_ver=2/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1:t-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1:t:raw_recreq_dt0] Path -> Partition: hdfs://bfdhadoopcool/warehouse/raw_recreq_dt0/l_ver=2/l_date=2015-07-28 Partition base file name: l_date=2015-07-28 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: l_date 2015-07-28 l_ver 2 properties: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim | columns customer,req_id,rec_type_code,number,uid,gid,fid,session_id,item_id,ip,method,num,flt,category,ref_page,bid,creation_time,version columns.comments columns.types string:string:int:int:string:string:string:string:string:string:string:int:string:string:string:string:string:string field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/raw_recreq_dt0/l_ver=2/l_date=2015-07-28 mapkey.delim , name default.raw_recreq_dt0 numFiles 235 numRows 0 partition_columns l_ver/l_date partition_columns.types string:string rawDataSize 0 serialization.ddl struct raw_recreq_dt0 { string customer, string req_id, i32 rec_type_code, i32 number, string uid, string gid, string fid, string session_id, string item_id, string ip, string method, i32 num, string flt, string category, string ref_page, string bid, string creation_time, string version} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format totalSize 30080297905 transient_lastDdlTime 1438106369 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 colelction.delim | columns customer,req_id,rec_type_code,number,uid,gid,fid,session_id,item_id,ip,method,num,flt,category,ref_page,bid,creation_time,version columns.comments columns.types string:string:int:int:string:string:string:string:string:string:string:int:string:string:string:string:string:string field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/raw_recreq_dt0 mapkey.delim , name default.raw_recreq_dt0 partition_columns l_ver/l_date partition_columns.types string:string serialization.ddl struct raw_recreq_dt0 { string customer, string req_id, i32 rec_type_code, i32 number, string uid, string gid, string fid, string session_id, string item_id, string ip, string method, i32 num, string flt, string category, string ref_page, string bid, string creation_time, string version} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.raw_recreq_dt0 name: default.raw_recreq_dt0 Truncated Path -> Alias: /raw_recreq_dt0/l_ver=2/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1:t-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1:t:raw_recreq_dt0] Needs Tagging: false Reduce Operator Tree: Group By Operator aggregations: count(DISTINCT KEY._col2:0._col0), count(VALUE._col1) keys: KEY._col0 (type: string), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 25066915 Data size: 7520074551 Basic stats: COMPLETE Column stats: NONE Filter Operator isSamplingPred: false predicate: (_col2 <= 3) (type: boolean) Statistics: Num rows: 8355638 Data size: 2506691416 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 8355638 Data size: 2506691416 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT _col1), sum(_col3) keys: _col0 (type: string) mode: complete outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 4177819 Data size: 1253345708 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: bigint), _col1 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), _col2 (type: bigint), _col2 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 4177819 Data size: 1253345708 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: true GlobalTableId: 0 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10007 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false Stage: Stage-10 Map Reduce Map Operator Tree: TableScan alias: raw_kafka_event_order_dt0 Statistics: Num rows: 143114 Data size: 28622892 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: customer is not null (type: boolean) Statistics: Num rows: 71557 Data size: 14311446 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: customer (type: string), session_id (type: string) outputColumnNames: customer, session_id Statistics: Num rows: 71557 Data size: 14311446 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT session_id) keys: customer (type: string), session_id (type: string) mode: hash outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 71557 Data size: 14311446 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 71557 Data size: 14311446 Basic stats: COMPLETE Column stats: NONE tag: -1 auto parallelism: false Path -> Alias: hdfs://bfdhadoopcool/warehouse/raw_kafka_event_order_dt0/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:raw_kafka_event_order_dt0] Path -> Partition: hdfs://bfdhadoopcool/warehouse/raw_kafka_event_order_dt0/l_date=2015-07-28 Partition base file name: l_date=2015-07-28 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: l_date 2015-07-28 properties: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim | columns customer,uid,session_id,gid,ip_address,event_type_id,ref_page,creation_time,tma,tmc,tmd,order_id,item_id,price,quantity,total,address,express,pay,mobile,consignee,epr columns.comments columns.types string:string:string:string:string:string:string:string:string:string:string:string:string:double:double:double:string:string:string:string:string:string field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/raw_kafka_event_order_dt0/l_date=2015-07-28 mapkey.delim , name default.raw_kafka_event_order_dt0 numFiles 1 numRows 0 partition_columns l_date partition_columns.types string rawDataSize 0 serialization.ddl struct raw_kafka_event_order_dt0 { string customer, string uid, string session_id, string gid, string ip_address, string event_type_id, string ref_page, string creation_time, string tma, string tmc, string tmd, string order_id, string item_id, double price, double quantity, double total, string address, string express, string pay, string mobile, string consignee, string epr} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format totalSize 28622892 transient_lastDdlTime 1438108224 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 colelction.delim | columns customer,uid,session_id,gid,ip_address,event_type_id,ref_page,creation_time,tma,tmc,tmd,order_id,item_id,price,quantity,total,address,express,pay,mobile,consignee,epr columns.comments columns.types string:string:string:string:string:string:string:string:string:string:string:string:string:double:double:double:string:string:string:string:string:string field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/raw_kafka_event_order_dt0 mapkey.delim , name default.raw_kafka_event_order_dt0 partition_columns l_date partition_columns.types string serialization.ddl struct raw_kafka_event_order_dt0 { string customer, string uid, string session_id, string gid, string ip_address, string event_type_id, string ref_page, string creation_time, string tma, string tmc, string tmd, string order_id, string item_id, double price, double quantity, double total, string address, string express, string pay, string mobile, string consignee, string epr} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.raw_kafka_event_order_dt0 name: default.raw_kafka_event_order_dt0 Truncated Path -> Alias: /raw_kafka_event_order_dt0/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:raw_kafka_event_order_dt0] Needs Tagging: false Reduce Operator Tree: Group By Operator aggregations: count(DISTINCT KEY._col1:0._col0) keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 35778 Data size: 7155622 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), _col1 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 35778 Data size: 7155622 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: true GlobalTableId: 0 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10008 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false Stage: Stage-11 Map Reduce Map Operator Tree: TableScan alias: items_rec Statistics: Num rows: 58521 Data size: 18258712 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: ((rec_type = 3) and customer is not null) (type: boolean) Statistics: Num rows: 14630 Data size: 4564599 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: customer (type: string), item_id (type: string), order_id (type: string), quantity (type: double) outputColumnNames: customer, item_id, order_id, quantity Statistics: Num rows: 14630 Data size: 4564599 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT item_id), sum(quantity), count(DISTINCT order_id, item_id) keys: customer (type: string), item_id (type: string), order_id (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 Statistics: Num rows: 14630 Data size: 4564599 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 14630 Data size: 4564599 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col4 (type: double) auto parallelism: false Path -> Alias: hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery2:items_rec] Path -> Partition: hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0/l_date=2015-07-28 Partition base file name: l_date=2015-07-28 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: l_date 2015-07-28 properties: COLUMN_STATS_ACCURATE true bucket_count -1 columns customer,creation_time,user_id,gid,session_id,fid,ip_address,ref_page,item_id,price,quantity,order_id,total_price,version,rec_type columns.comments columns.types string:string:string:string:string:string:string:string:string:double:double:string:double:string:int field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0/l_date=2015-07-28 name default.bi_order_detail_rec_dt0 numFiles 2 numRows 0 partition_columns l_date partition_columns.types string rawDataSize 0 serialization.ddl struct bi_order_detail_rec_dt0 { string customer, string creation_time, string user_id, string gid, string session_id, string fid, string ip_address, string ref_page, string item_id, double price, double quantity, string order_id, double total_price, string version, i32 rec_type} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format totalSize 18258712 transient_lastDdlTime 1438118346 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns customer,creation_time,user_id,gid,session_id,fid,ip_address,ref_page,item_id,price,quantity,order_id,total_price,version,rec_type columns.comments columns.types string:string:string:string:string:string:string:string:string:double:double:string:double:string:int field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0 name default.bi_order_detail_rec_dt0 partition_columns l_date partition_columns.types string serialization.ddl struct bi_order_detail_rec_dt0 { string customer, string creation_time, string user_id, string gid, string session_id, string fid, string ip_address, string ref_page, string item_id, double price, double quantity, string order_id, double total_price, string version, i32 rec_type} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bi_order_detail_rec_dt0 name: default.bi_order_detail_rec_dt0 Truncated Path -> Alias: /bi_order_detail_rec_dt0/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery2:items_rec] Needs Tagging: false Reduce Operator Tree: Group By Operator aggregations: count(DISTINCT KEY._col1:0._col0), sum(VALUE._col1), count(DISTINCT KEY._col1:1._col0, KEY._col1:1._col1) keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 7315 Data size: 2282299 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), _col1 (type: bigint), 0.0 (type: double), 0 (type: bigint), UDFToLong(_col2) (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), _col3 (type: bigint), 0 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 7315 Data size: 2282299 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: true GlobalTableId: 0 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10009 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false Stage: Stage-12 Map Reduce Map Operator Tree: TableScan alias: raw_recresult_dt1 Statistics: Num rows: 3694316260 Data size: 738863252034 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: customer is not null (type: boolean) Statistics: Num rows: 1847158130 Data size: 369431626017 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: customer (type: string), rec_item_id (type: string) outputColumnNames: customer, rec_item_id Statistics: Num rows: 1847158130 Data size: 369431626017 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(rec_item_id), count(DISTINCT rec_item_id) keys: customer (type: string), rec_item_id (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1847158130 Data size: 369431626017 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 1847158130 Data size: 369431626017 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col2 (type: bigint) auto parallelism: false Path -> Alias: hdfs://bfdhadoopcool/warehouse/raw_recresult_dt1/l_date=2015-07-28 [null-subquery2:t-subquery2:raw_recresult_dt1] Path -> Partition: hdfs://bfdhadoopcool/warehouse/raw_recresult_dt1/l_date=2015-07-28 Partition base file name: l_date=2015-07-28 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: l_date 2015-07-28 properties: COLUMN_STATS_ACCURATE true bucket_count -1 columns customer,rec_item_id,req_id,creation_time,uuid,abtest,bid,extend,rec_customer columns.comments columns.types string:string:string:string:string:string:string:map:string field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/raw_recresult_dt1/l_date=2015-07-28 mapkey.delim , name default.raw_recresult_dt1 numFiles 3532 numRows 0 partition_columns l_date partition_columns.types string rawDataSize 0 serialization.ddl struct raw_recresult_dt1 { string customer, string rec_item_id, string req_id, string creation_time, string uuid, string abtest, string bid, map extend, string rec_customer} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format totalSize 738863252034 transient_lastDdlTime 1438115762 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns customer,rec_item_id,req_id,creation_time,uuid,abtest,bid,extend,rec_customer columns.comments columns.types string:string:string:string:string:string:string:map:string field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/raw_recresult_dt1 mapkey.delim , name default.raw_recresult_dt1 partition_columns l_date partition_columns.types string serialization.ddl struct raw_recresult_dt1 { string customer, string rec_item_id, string req_id, string creation_time, string uuid, string abtest, string bid, map extend, string rec_customer} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.raw_recresult_dt1 name: default.raw_recresult_dt1 Truncated Path -> Alias: /raw_recresult_dt1/l_date=2015-07-28 [null-subquery2:t-subquery2:raw_recresult_dt1] Needs Tagging: false Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0), count(DISTINCT KEY._col1:0._col0) keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 923579065 Data size: 184715813008 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), _col1 (type: bigint), _col2 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 923579065 Data size: 184715813008 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: true GlobalTableId: 0 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10010 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false Stage: Stage-13 Map Reduce Map Operator Tree: TableScan alias: bi_order_detail_rec_dt0 Statistics: Num rows: 89503 Data size: 18258712 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: ((rec_type = 9) and customer is not null) (type: boolean) Statistics: Num rows: 22376 Data size: 4564729 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: customer (type: string), item_id (type: string) outputColumnNames: customer, item_id Statistics: Num rows: 22376 Data size: 4564729 Basic stats: COMPLETE Column stats: NONE Group By Operator keys: customer (type: string), item_id (type: string) mode: hash outputColumnNames: _col0, _col1 Statistics: Num rows: 22376 Data size: 4564729 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) Statistics: Num rows: 22376 Data size: 4564729 Basic stats: COMPLETE Column stats: NONE tag: -1 auto parallelism: false Path -> Alias: hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0/l_date=2015-07-28 [null-subquery1-subquery2:t-subquery1-subquery2:created_items_by_rec:bi_order_detail_rec_dt0] Path -> Partition: hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0/l_date=2015-07-28 Partition base file name: l_date=2015-07-28 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: l_date 2015-07-28 properties: COLUMN_STATS_ACCURATE true bucket_count -1 columns customer,creation_time,user_id,gid,session_id,fid,ip_address,ref_page,item_id,price,quantity,order_id,total_price,version,rec_type columns.comments columns.types string:string:string:string:string:string:string:string:string:double:double:string:double:string:int field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0/l_date=2015-07-28 name default.bi_order_detail_rec_dt0 numFiles 2 numRows 0 partition_columns l_date partition_columns.types string rawDataSize 0 serialization.ddl struct bi_order_detail_rec_dt0 { string customer, string creation_time, string user_id, string gid, string session_id, string fid, string ip_address, string ref_page, string item_id, double price, double quantity, string order_id, double total_price, string version, i32 rec_type} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format totalSize 18258712 transient_lastDdlTime 1438118346 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns customer,creation_time,user_id,gid,session_id,fid,ip_address,ref_page,item_id,price,quantity,order_id,total_price,version,rec_type columns.comments columns.types string:string:string:string:string:string:string:string:string:double:double:string:double:string:int field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0 name default.bi_order_detail_rec_dt0 partition_columns l_date partition_columns.types string serialization.ddl struct bi_order_detail_rec_dt0 { string customer, string creation_time, string user_id, string gid, string session_id, string fid, string ip_address, string ref_page, string item_id, double price, double quantity, string order_id, double total_price, string version, i32 rec_type} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bi_order_detail_rec_dt0 name: default.bi_order_detail_rec_dt0 Truncated Path -> Alias: /bi_order_detail_rec_dt0/l_date=2015-07-28 [null-subquery1-subquery2:t-subquery1-subquery2:created_items_by_rec:bi_order_detail_rec_dt0] Needs Tagging: false Reduce Operator Tree: Group By Operator keys: KEY._col0 (type: string), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 11188 Data size: 2282364 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 11188 Data size: 2282364 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: true GlobalTableId: 0 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10011 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1 columns.types string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false Stage: Stage-14 Map Reduce Map Operator Tree: TableScan alias: items_rec Statistics: Num rows: 58521 Data size: 18258712 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: ((rec_type >= 6) and customer is not null) (type: boolean) Statistics: Num rows: 9754 Data size: 3043274 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: customer (type: string), item_id (type: string), order_id (type: string), rec_type (type: int), quantity (type: double) outputColumnNames: customer, item_id, order_id, rec_type, quantity Statistics: Num rows: 9754 Data size: 3043274 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT item_id), sum(quantity), count(DISTINCT order_id, item_id), count(DISTINCT CASE (rec_type) WHEN (9) THEN (item_id) ELSE (null) END) keys: customer (type: string), item_id (type: string), order_id (type: string), CASE (rec_type) WHEN (9) THEN (item_id) ELSE (null) END (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 Statistics: Num rows: 9754 Data size: 3043274 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) sort order: ++++ Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 9754 Data size: 3043274 Basic stats: COMPLETE Column stats: NONE tag: -1 value expressions: _col5 (type: double) auto parallelism: false Path -> Alias: hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery1-subquery2:items_rec] Path -> Partition: hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0/l_date=2015-07-28 Partition base file name: l_date=2015-07-28 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: l_date 2015-07-28 properties: COLUMN_STATS_ACCURATE true bucket_count -1 columns customer,creation_time,user_id,gid,session_id,fid,ip_address,ref_page,item_id,price,quantity,order_id,total_price,version,rec_type columns.comments columns.types string:string:string:string:string:string:string:string:string:double:double:string:double:string:int field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0/l_date=2015-07-28 name default.bi_order_detail_rec_dt0 numFiles 2 numRows 0 partition_columns l_date partition_columns.types string rawDataSize 0 serialization.ddl struct bi_order_detail_rec_dt0 { string customer, string creation_time, string user_id, string gid, string session_id, string fid, string ip_address, string ref_page, string item_id, double price, double quantity, string order_id, double total_price, string version, i32 rec_type} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format totalSize 18258712 transient_lastDdlTime 1438118346 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 columns customer,creation_time,user_id,gid,session_id,fid,ip_address,ref_page,item_id,price,quantity,order_id,total_price,version,rec_type columns.comments columns.types string:string:string:string:string:string:string:string:string:double:double:string:double:string:int field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/bi_order_detail_rec_dt0 name default.bi_order_detail_rec_dt0 partition_columns l_date partition_columns.types string serialization.ddl struct bi_order_detail_rec_dt0 { string customer, string creation_time, string user_id, string gid, string session_id, string fid, string ip_address, string ref_page, string item_id, double price, double quantity, string order_id, double total_price, string version, i32 rec_type} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bi_order_detail_rec_dt0 name: default.bi_order_detail_rec_dt0 Truncated Path -> Alias: /bi_order_detail_rec_dt0/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery1-subquery2:items_rec] Needs Tagging: false Reduce Operator Tree: Group By Operator aggregations: count(DISTINCT KEY._col1:0._col0), sum(VALUE._col1), count(DISTINCT KEY._col1:1._col0, KEY._col1:1._col1), count(DISTINCT KEY._col1:2._col0) keys: KEY._col0 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3, _col4 Statistics: Num rows: 4877 Data size: 1521637 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), _col1 (type: bigint), 0.0 (type: double), 0 (type: bigint), UDFToLong(_col2) (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), _col3 (type: bigint), 0 (type: bigint), _col4 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 4877 Data size: 1521637 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: true GlobalTableId: 0 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10012 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false Stage: Stage-15 Map Reduce Map Operator Tree: TableScan alias: raw_recreq_dt0 Statistics: Num rows: 100267659 Data size: 30080297905 Basic stats: COMPLETE Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: customer is not null (type: boolean) Statistics: Num rows: 50133830 Data size: 15040149102 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: customer (type: string), gid (type: string), session_id (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 50133830 Data size: 15040149102 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT _col1) keys: _col0 (type: string), _col2 (type: string), _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 50133830 Data size: 15040149102 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) Statistics: Num rows: 50133830 Data size: 15040149102 Basic stats: COMPLETE Column stats: NONE tag: -1 auto parallelism: false Path -> Alias: hdfs://bfdhadoopcool/warehouse/raw_recreq_dt0/l_ver=2/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery2:t-subquery2:t2:raw_recreq_dt0] Path -> Partition: hdfs://bfdhadoopcool/warehouse/raw_recreq_dt0/l_ver=2/l_date=2015-07-28 Partition base file name: l_date=2015-07-28 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: l_date 2015-07-28 l_ver 2 properties: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim | columns customer,req_id,rec_type_code,number,uid,gid,fid,session_id,item_id,ip,method,num,flt,category,ref_page,bid,creation_time,version columns.comments columns.types string:string:int:int:string:string:string:string:string:string:string:int:string:string:string:string:string:string field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/raw_recreq_dt0/l_ver=2/l_date=2015-07-28 mapkey.delim , name default.raw_recreq_dt0 numFiles 235 numRows 0 partition_columns l_ver/l_date partition_columns.types string:string rawDataSize 0 serialization.ddl struct raw_recreq_dt0 { string customer, string req_id, i32 rec_type_code, i32 number, string uid, string gid, string fid, string session_id, string item_id, string ip, string method, i32 num, string flt, string category, string ref_page, string bid, string creation_time, string version} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format totalSize 30080297905 transient_lastDdlTime 1438106369 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 colelction.delim | columns customer,req_id,rec_type_code,number,uid,gid,fid,session_id,item_id,ip,method,num,flt,category,ref_page,bid,creation_time,version columns.comments columns.types string:string:int:int:string:string:string:string:string:string:string:int:string:string:string:string:string:string field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/raw_recreq_dt0 mapkey.delim , name default.raw_recreq_dt0 partition_columns l_ver/l_date partition_columns.types string:string serialization.ddl struct raw_recreq_dt0 { string customer, string req_id, i32 rec_type_code, i32 number, string uid, string gid, string fid, string session_id, string item_id, string ip, string method, i32 num, string flt, string category, string ref_page, string bid, string creation_time, string version} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.raw_recreq_dt0 name: default.raw_recreq_dt0 Truncated Path -> Alias: /raw_recreq_dt0/l_ver=2/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery2:t-subquery2:t2:raw_recreq_dt0] Needs Tagging: false Reduce Operator Tree: Group By Operator aggregations: count(DISTINCT KEY._col2:0._col0) keys: KEY._col0 (type: string), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 25066915 Data size: 7520074551 Basic stats: COMPLETE Column stats: NONE Filter Operator isSamplingPred: false predicate: (_col2 <= 3) (type: boolean) Statistics: Num rows: 8355638 Data size: 2506691416 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 8355638 Data size: 2506691416 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: true GlobalTableId: 0 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10013 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1 columns.types string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false Stage: Stage-16 Map Reduce Map Operator Tree: TableScan GatherStats: false Union Statistics: Num rows: 21576952 Data size: 9117348420 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 21576952 Data size: 9117348420 Basic stats: COMPLETE Column stats: NONE Group By Operator keys: _col0 (type: string), _col1 (type: string) mode: hash outputColumnNames: _col0, _col1 Statistics: Num rows: 21576952 Data size: 9117348420 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 21576952 Data size: 9117348420 Basic stats: COMPLETE Column stats: NONE tag: -1 auto parallelism: false TableScan GatherStats: false Union Statistics: Num rows: 21576952 Data size: 9117348420 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 21576952 Data size: 9117348420 Basic stats: COMPLETE Column stats: NONE Group By Operator keys: _col0 (type: string), _col1 (type: string) mode: hash outputColumnNames: _col0, _col1 Statistics: Num rows: 21576952 Data size: 9117348420 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string) sort order: ++ Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 21576952 Data size: 9117348420 Basic stats: COMPLETE Column stats: NONE tag: -1 auto parallelism: false Path -> Alias: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10013 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10013] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10015 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10015] Path -> Partition: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10013 Partition base file name: -mr-10013 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1 columns.types string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1 columns.types string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10015 Partition base file name: -mr-10015 input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1 columns.types string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1 columns.types string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe Truncated Path -> Alias: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10013 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10013] hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10015 [hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10015] Needs Tagging: false Reduce Operator Tree: Group By Operator keys: KEY._col0 (type: string), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 10788476 Data size: 4558674210 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 10788476 Data size: 4558674210 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT _col1) keys: _col0 (type: string) mode: complete outputColumnNames: _col0, _col1 Statistics: Num rows: 5394238 Data size: 2279337105 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), 0 (type: bigint), 0 (type: bigint), _col1 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0.0 (type: double), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint), 0 (type: bigint) outputColumnNames: _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28 Statistics: Num rows: 5394238 Data size: 2279337105 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: true GlobalTableId: 0 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10014 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col1,_col2,_col3,_col4,_col5,_col6,_col7,_col8,_col9,_col10,_col11,_col12,_col13,_col14,_col15,_col16,_col17,_col18,_col19,_col20,_col21,_col22,_col23,_col24,_col25,_col26,_col27,_col28 columns.types string,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint,bigint,bigint,bigint,double,bigint,bigint,bigint,bigint,bigint escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false Stage: Stage-17 Map Reduce Map Operator Tree: TableScan alias: raw_event_dt0 Statistics: Num rows: 105770516 Data size: 52885258037 Basic stats: PARTIAL Column stats: NONE GatherStats: false Filter Operator isSamplingPred: false predicate: (((version = '2') or (event_type_id <> 'PageView')) and customer is not null) (type: boolean) Statistics: Num rows: 79327887 Data size: 39663943527 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: customer (type: string), gid (type: string), session_id (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 79327887 Data size: 39663943527 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count(DISTINCT _col1) keys: _col0 (type: string), _col2 (type: string), _col1 (type: string) mode: hash outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 79327887 Data size: 39663943527 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) sort order: +++ Map-reduce partition columns: _col0 (type: string), _col1 (type: string) Statistics: Num rows: 79327887 Data size: 39663943527 Basic stats: COMPLETE Column stats: NONE tag: -1 auto parallelism: false Path -> Alias: hdfs://bfdhadoopcool/warehouse/raw_event_dt0/l_ver=0/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1:t-subquery1:t1:raw_event_dt0] hdfs://bfdhadoopcool/warehouse/raw_event_dt0/l_ver=1/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1:t-subquery1:t1:raw_event_dt0] hdfs://bfdhadoopcool/warehouse/raw_event_dt0/l_ver=2/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1:t-subquery1:t1:raw_event_dt0] Path -> Partition: hdfs://bfdhadoopcool/warehouse/raw_event_dt0/l_ver=0/l_date=2015-07-28 Partition base file name: l_date=2015-07-28 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: l_date 2015-07-28 l_ver 0 properties: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim | columns customer,item_id,user_id,gid,session_id,event_type_id,ip_address,ref_page,creation_time,price,quantity,order_id,total_price,req_id,fid,cat_id,bid,num,version,page_type,page_url columns.comments columns.types string:string:string:string:string:string:string:string:string:double:double:string:double:string:string:string:string:bigint:string:string:string field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/raw_event_dt0/l_ver=0/l_date=2015-07-28 mapkey.delim , name default.raw_event_dt0 numFiles 1 numRows 0 partition_columns l_ver/l_date partition_columns.types string:string rawDataSize 0 serialization.ddl struct raw_event_dt0 { string customer, string item_id, string user_id, string gid, string session_id, string event_type_id, string ip_address, string ref_page, string creation_time, double price, double quantity, string order_id, double total_price, string req_id, string fid, string cat_id, string bid, i64 num, string version, string page_type, string page_url} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format totalSize 0 transient_lastDdlTime 1438115488 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 colelction.delim | columns customer,item_id,user_id,gid,session_id,event_type_id,ip_address,ref_page,creation_time,price,quantity,order_id,total_price,req_id,fid,cat_id,bid,num,version,page_type,page_url columns.comments columns.types string:string:string:string:string:string:string:string:string:double:double:string:double:string:string:string:string:bigint:string:string:string field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/raw_event_dt0 mapkey.delim , name default.raw_event_dt0 partition_columns l_ver/l_date partition_columns.types string:string serialization.ddl struct raw_event_dt0 { string customer, string item_id, string user_id, string gid, string session_id, string event_type_id, string ip_address, string ref_page, string creation_time, double price, double quantity, string order_id, double total_price, string req_id, string fid, string cat_id, string bid, i64 num, string version, string page_type, string page_url} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.raw_event_dt0 name: default.raw_event_dt0 hdfs://bfdhadoopcool/warehouse/raw_event_dt0/l_ver=1/l_date=2015-07-28 Partition base file name: l_date=2015-07-28 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: l_date 2015-07-28 l_ver 1 properties: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim | columns customer,item_id,user_id,gid,session_id,event_type_id,ip_address,ref_page,creation_time,price,quantity,order_id,total_price,req_id,fid,cat_id,bid,num,version,page_type,page_url columns.comments columns.types string:string:string:string:string:string:string:string:string:double:double:string:double:string:string:string:string:bigint:string:string:string field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/raw_event_dt0/l_ver=1/l_date=2015-07-28 mapkey.delim , name default.raw_event_dt0 numFiles 1 numRows 0 partition_columns l_ver/l_date partition_columns.types string:string rawDataSize 0 serialization.ddl struct raw_event_dt0 { string customer, string item_id, string user_id, string gid, string session_id, string event_type_id, string ip_address, string ref_page, string creation_time, double price, double quantity, string order_id, double total_price, string req_id, string fid, string cat_id, string bid, i64 num, string version, string page_type, string page_url} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format totalSize 0 transient_lastDdlTime 1438115327 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 colelction.delim | columns customer,item_id,user_id,gid,session_id,event_type_id,ip_address,ref_page,creation_time,price,quantity,order_id,total_price,req_id,fid,cat_id,bid,num,version,page_type,page_url columns.comments columns.types string:string:string:string:string:string:string:string:string:double:double:string:double:string:string:string:string:bigint:string:string:string field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/raw_event_dt0 mapkey.delim , name default.raw_event_dt0 partition_columns l_ver/l_date partition_columns.types string:string serialization.ddl struct raw_event_dt0 { string customer, string item_id, string user_id, string gid, string session_id, string event_type_id, string ip_address, string ref_page, string creation_time, double price, double quantity, string order_id, double total_price, string req_id, string fid, string cat_id, string bid, i64 num, string version, string page_type, string page_url} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.raw_event_dt0 name: default.raw_event_dt0 hdfs://bfdhadoopcool/warehouse/raw_event_dt0/l_ver=2/l_date=2015-07-28 Partition base file name: l_date=2015-07-28 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: l_date 2015-07-28 l_ver 2 properties: COLUMN_STATS_ACCURATE true bucket_count -1 colelction.delim | columns customer,item_id,user_id,gid,session_id,event_type_id,ip_address,ref_page,creation_time,price,quantity,order_id,total_price,req_id,fid,cat_id,bid,num,version,page_type,page_url columns.comments columns.types string:string:string:string:string:string:string:string:string:double:double:string:double:string:string:string:string:bigint:string:string:string field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/raw_event_dt0/l_ver=2/l_date=2015-07-28 mapkey.delim , name default.raw_event_dt0 numFiles 1035 numRows 0 partition_columns l_ver/l_date partition_columns.types string:string rawDataSize 0 serialization.ddl struct raw_event_dt0 { string customer, string item_id, string user_id, string gid, string session_id, string event_type_id, string ip_address, string ref_page, string creation_time, double price, double quantity, string order_id, double total_price, string req_id, string fid, string cat_id, string bid, i64 num, string version, string page_type, string page_url} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format totalSize 52885258037 transient_lastDdlTime 1438115228 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 colelction.delim | columns customer,item_id,user_id,gid,session_id,event_type_id,ip_address,ref_page,creation_time,price,quantity,order_id,total_price,req_id,fid,cat_id,bid,num,version,page_type,page_url columns.comments columns.types string:string:string:string:string:string:string:string:string:double:double:string:double:string:string:string:string:bigint:string:string:string field.delim file.inputformat org.apache.hadoop.mapred.TextInputFormat file.outputformat org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat location hdfs://bfdhadoopcool/warehouse/raw_event_dt0 mapkey.delim , name default.raw_event_dt0 partition_columns l_ver/l_date partition_columns.types string:string serialization.ddl struct raw_event_dt0 { string customer, string item_id, string user_id, string gid, string session_id, string event_type_id, string ip_address, string ref_page, string creation_time, double price, double quantity, string order_id, double total_price, string req_id, string fid, string cat_id, string bid, i64 num, string version, string page_type, string page_url} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe serialization.null.format serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.raw_event_dt0 name: default.raw_event_dt0 Truncated Path -> Alias: /raw_event_dt0/l_ver=0/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1:t-subquery1:t1:raw_event_dt0] /raw_event_dt0/l_ver=1/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1:t-subquery1:t1:raw_event_dt0] /raw_event_dt0/l_ver=2/l_date=2015-07-28 [null-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery1-subquery2:t-subquery1:t-subquery1:t1:raw_event_dt0] Needs Tagging: false Reduce Operator Tree: Group By Operator aggregations: count(DISTINCT KEY._col2:0._col0) keys: KEY._col0 (type: string), KEY._col1 (type: string) mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 39663943 Data size: 19831971513 Basic stats: COMPLETE Column stats: NONE Filter Operator isSamplingPred: false predicate: (_col2 <= 3) (type: boolean) Statistics: Num rows: 13221314 Data size: 6610657004 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 13221314 Data size: 6610657004 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: true GlobalTableId: 0 directory: hdfs://bfdhadoopcool/tmp/hive/hadoop/1d75fc01-eaeb-4b29-9283-87d132c97844/hive_2015-07-29_14-27-51_122_6788669498397282851-1/-mr-10015 NumFilesPerFileSink: 1 table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: columns _col0,_col1 columns.types string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false