Uploaded image for project: 'Hive'
  1. Hive
  2. HIVE-5393

Hive Multi insert statement is not logging how many rows loaded for individual insert statements

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Open
    • Minor
    • Resolution: Unresolved
    • 0.9.0
    • None
    • Diagnosability
    • None
    • Hive 0.9.0
      Cloudera CDH3U4

    Description

      When I am running multi insert statment in Hive to copy data from a single source table to a target table but into two different partitions, The log is not showing how many rows got loaded into two different partitions by the separate insert statements. It is only showing how many rows loaded for only insert statement not both of them.

      Below is the query:

      =============================================================
      from WBUSINESS_STR
      insert OVERWRITE table WBUSINESS_intraday_30mins PARTITION (data_year='2013',data_month='09',data_day='23',data_hourmin='0000',data_run_id='201309230001')
      select
      year(DB_COMMIT_TIMESTAMP ) as year,
      month(DB_COMMIT_TIMESTAMP) as month,
      day(DB_COMMIT_TIMESTAMP) as day,
      concat(hour(DB_COMMIT_TIMESTAMP),minute(DB_COMMIT_TIMESTAMP)) as hourmin,
      '201309230001' as RUN_ID,
      cast(DB_COMMIT_TIMESTAMP as timestamp),
      DB_ACTION_TYPE,
      cast( RBA as bigint),
      cast( ID as bigint ),
      cast( ACCOUNT_NUMBER as bigint ),
      cast( TIME_CREATED as bigint ),
      cast( FLAGS as bigint ),
      NAME ,
      NAME_UPPER ,
      cast( ADDRESS_ID as bigint ),
      cast( EIN as bigint ),
      DUNS_NUMBER ,
      cast( SIC as bigint ),
      cast( TYPE as bigint ),
      STATE_INC ,
      cast( AVERAGE_TICKET as bigint ),
      SERVICE_EMAIL ,
      SERVICE_PHONE ,
      cast(concat(substr(TIME_ROW_UPDATED,1,10),' ',substr(TIME_ROW_UPDATED,12)) as timestamp),
      DISPUTE_EMAIL ,
      DOING_BUSINESS_AS
      where
      cast(DB_COMMIT_TIMESTAMP as timestamp) >= cast('2013-09-23 00:00:00' as timestamp)
      and cast(DB_COMMIT_TIMESTAMP as timestamp) < cast('2013-09-23 00:30:00' as timestamp)
      insert OVERWRITE table WBUSINESS_intraday_30mins PARTITION (data_year='2013',data_month='09',data_day='22',data_hourmin='2330',data_run_id='201309230001')
      select
      year(DB_COMMIT_TIMESTAMP ) as year,
      month(DB_COMMIT_TIMESTAMP) as month,
      day(DB_COMMIT_TIMESTAMP) as day,
      concat(hour(DB_COMMIT_TIMESTAMP),minute(DB_COMMIT_TIMESTAMP)) as hourmin,
      '201309230001' as RUN_ID,
      cast(DB_COMMIT_TIMESTAMP as timestamp),
      DB_ACTION_TYPE,
      cast( RBA as bigint),
      cast( ID as bigint ),
      cast( ACCOUNT_NUMBER as bigint ),
      cast( TIME_CREATED as bigint ),
      cast( FLAGS as bigint ),
      NAME ,
      NAME_UPPER ,
      cast( ADDRESS_ID as bigint ),
      cast( EIN as bigint ),
      DUNS_NUMBER ,
      cast( SIC as bigint ),
      cast( TYPE as bigint ),
      STATE_INC ,
      cast( AVERAGE_TICKET as bigint ),
      SERVICE_EMAIL ,
      SERVICE_PHONE ,
      cast(concat(substr(TIME_ROW_UPDATED,1,10),' ',substr(TIME_ROW_UPDATED,12)) as timestamp),
      DISPUTE_EMAIL ,
      DOING_BUSINESS_AS
      where
      cast(DB_COMMIT_TIMESTAMP as timestamp) >= cast('2013-09-22 23:30:00' as timestamp)
      and cast(DB_COMMIT_TIMESTAMP as timestamp) < cast('2013-09-23 00:00:00' as timestamp)
      ============================================

      The Hive log says:

      Total MapReduce jobs = 1
      Launching Job 1 out of 1
      Number of reduce tasks is set to 0 since there's no reduce operator
      Starting Job = job_201309261542_0042, Tracking URL = http://dmdevetllvs01:50030/jobdetails.jsp?jobid=job_201309261542_0042
      Kill Command = /dmdev/data01/paypal/HADOOP/node1/hadoop-0.20.2-cdh3u4/bin/hadoop job -Dmapred.job.tracker=dmdevetllvs01:9010 -kill job_201309261542_0042
      Hadoop job information for Stage-2: number of mappers: 1; number of reducers: 0
      2013-09-27 16:24:26,542 Stage-2 map = 0%, reduce = 0%
      2013-09-27 16:24:28,565 Stage-2 map = 100%, reduce = 0%, Cumulative CPU 2.52 sec
      2013-09-27 16:24:29,574 Stage-2 map = 100%, reduce = 100%, Cumulative CPU 2.52 sec
      MapReduce Total cumulative CPU time: 2 seconds 520 msec
      Ended Job = job_201309261542_0042
      Ended Job = 730112486, job is filtered out (removed at runtime).
      Ended Job = -507348248, job is filtered out (removed at runtime).
      Moving data to: hdfs://dmdevetllvs01.qa.paypal.com/tmp/hive-sanddas/hive_2013-09-27_16-24-19_784_1282210997007172719/-ext-10000
      Moving data to: hdfs://dmdevetllvs01.qa.paypal.com/tmp/hive-sanddas/hive_2013-09-27_16-24-19_784_1282210997007172719/-ext-10002
      Loading data to table default.wbusiness_intraday_30mins partition (data_year=2013, data_month=09, data_day=22, data_hourmin=2330, data_run_id=201309230001)
      Deleted hdfs://dmdevetllvs01.qa.paypal.com/sys/pp_dm/sanddas/import/site/CONF/WBUSINESS/RCFILE/incremental/data_year=2013/data_month=09/data_day=22/data_hourmin=2330/data_run_id=201309230001
      Partition default.wbusiness_intraday_30mins

      {data_year=2013, data_month=09, data_day=22, data_hourmin=2330, data_run_id=201309230001}

      stats: [num_files: 1, num_rows: 0, total_size: 12283, raw_data_size: 0]
      Table default.wbusiness_intraday_30mins stats: [num_partitions: 7, num_files: 7, num_rows: 0, total_size: 62018, raw_data_size: 0]
      Loading data to table default.wbusiness_intraday_30mins partition (data_year=2013, data_month=09, data_day=23, data_hourmin=0000, data_run_id=201309230001)
      Deleted hdfs://dmdevetllvs01.qa.paypal.com/sys/pp_dm/sanddas/import/site/CONF/WBUSINESS/RCFILE/incremental/data_year=2013/data_month=09/data_day=23/data_hourmin=0000/data_run_id=201309230001
      Partition default.wbusiness_intraday_30mins

      {data_year=2013, data_month=09, data_day=23, data_hourmin=0000, data_run_id=201309230001}

      stats: [num_files: 1, num_rows: 0, total_size: 1870, raw_data_size: 0]
      Table default.wbusiness_intraday_30mins stats: [num_partitions: 7, num_files: 7, num_rows: 0, total_size: 62018, raw_data_size: 0]
      141 Rows loaded to wbusiness_intraday_30mins
      MapReduce Jobs Launched:
      Job 0: Map: 1 Cumulative CPU: 2.52 sec HDFS Read: 4191 HDFS Write: 14153 SUCCESS
      Total MapReduce CPU Time Spent: 2 seconds 520 msec
      OK
      Time taken: 10.771 seconds
      ====================================================================

      It is only showing 141 rows loaded for a partition.
      It is not showing another 21 rows loaded in another partition.

      The source table has 161 rows.

      Please let us know if this is normal.

      Attachments

        Activity

          People

            Unassigned Unassigned
            sanddas154 Sandip Das
            Votes:
            0 Vote for this issue
            Watchers:
            4 Start watching this issue

            Dates

              Created:
              Updated: