Details
-
Bug
-
Status: Resolved
-
Major
-
Resolution: Resolved
-
2.1.1
-
None
-
None
Description
SQL: insert *overwrite *table dzdz_fpxx_dzfp partition(nian) select * from test.dzdz_fpxx_dzfp
create table dzdz_fpxx_dzfp ( FPDM string, FPHM string, KPRQ timestamp, FPLY string) partitioned by (nian string) stored as parquet; create table test.dzdz_fpxx_dzfp ( FPDM string, FPHM string, KPRQ timestamp, FPLY string, nian string) stored as textfile location "/origin/data/dzfp_origin/"
The execute insert sql:
SQL: insert overwrite table dzdz_fpxx_dzfp partition(nian) select * from test.dzdz_fpxx_dzfp
INFO : Compiling command(queryId=hive_20200519100412_12f2b39c-45f4-4f4c-9261-32ca86fa28db): insert overwrite table dzdz_fpxx_dzfp partition(nian) select * from test.dzdz_fpxx_dzfp INFO : Semantic Analysis Completed INFO : Returning Hive schema: Schema(fieldSchemas:[FieldSchema(name:dzdz_fpxx_dzfp.fpdm, type:string, comment:null), FieldSchema(name:dzdz_fpxx_dzfp.fphm, type:string, comment:null), FieldSchema(name:dzdz_fpxx_dzfp.kprq, type:timestamp, comment:null), FieldSchema(name:dzdz_fpxx_dzfp.tspz_dm, type:string, comment:null), FieldSchema(name:dzdz_fpxx_dzfp.fpzt_bz, type:string, comment:null), FieldSchema(name:dzdz_fpxx_dzfp.fply, type:string, comment:null), FieldSchema(name:dzdz_fpxx_dzfp.nian, type:string, comment:null)], properties:null) INFO : Completed compiling command(queryId=hive_20200519100412_12f2b39c-45f4-4f4c-9261-32ca86fa28db); Time taken: 1.719 seconds INFO : Executing command(queryId=hive_20200519100412_12f2b39c-45f4-4f4c-9261-32ca86fa28db): insert overwrite table dzdz_fpxx_dzfp partition(nian) select * from test.dzdz_fpxx_dzfp WARN : INFO : Query ID = hive_20200519100412_12f2b39c-45f4-4f4c-9261-32ca86fa28db INFO : Total jobs = 3 INFO : Launching Job 1 out of 3 INFO : Starting task [Stage-1:MAPRED] in serial mode INFO : Number of reduce tasks is set to 0 since there's no reduce operator INFO : number of splits:1 INFO : Submitting tokens for job: job_1589451904439_0049 INFO : Executing with tokens: [] INFO : The url to track the job: http://qcj37.hde.com:8088/proxy/application_1589451904439_0049/ INFO : Starting Job = job_1589451904439_0049, Tracking URL = http://qcj37.hde.com:8088/proxy/application_1589451904439_0049/ INFO : Kill Command = /usr/hdp/current/hadoop-client/bin/hadoop job -kill job_1589451904439_0049 INFO : Hadoop job information for Stage-1: number of mappers: 1; number of reducers: 0 INFO : 2020-05-19 10:06:56,823 Stage-1 map = 0%, reduce = 0% INFO : 2020-05-19 10:07:06,317 Stage-1 map = 100%, reduce = 0%, Cumulative CPU 4.3 sec INFO : MapReduce Total cumulative CPU time: 4 seconds 300 msec INFO : Ended Job = job_1589451904439_0049 INFO : Starting task [Stage-7:CONDITIONAL] in serial mode INFO : Stage-4 is selected by condition resolver. INFO : Stage-3 is filtered out by condition resolver. INFO : Stage-5 is filtered out by condition resolver. INFO : Starting task [Stage-4:MOVE] in serial mode INFO : Moving data to directory hdfs://mycluster/warehouse/tablespace/managed/hive/dzdz_fpxx_dzfp/.hive-staging_hive_2020-05-19_10-04-12_468_7695595367555279265-3/-ext-10000 from hdfs://mycluster/warehouse/tablespace/managed/hive/dzdz_fpxx_dzfp/.hive-staging_hive_2020-05-19_10-04-12_468_7695595367555279265-3/-ext-10002 INFO : Starting task [Stage-0:MOVE] in serial mode INFO : Loading data to table default.dzdz_fpxx_dzfp partition (nian=null) from hdfs://mycluster/warehouse/tablespace/managed/hive/dzdz_fpxx_dzfp/.hive-staging_hive_2020-05-19_10-04-12_468_7695595367555279265-3/-ext-10000 INFO : ERROR : FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.MoveTask. Exception when loading 2 in table dzdz_fpxx_dzfp with loadPath=hdfs://mycluster/warehouse/tablespace/managed/hive/dzdz_fpxx_dzfp/.hive-staging_hive_2020-05-19_10-04-12_468_7695595367555279265-3/-ext-10000 INFO : MapReduce Jobs Launched: INFO : Stage-Stage-1: Map: 1 Cumulative CPU: 4.3 sec HDFS Read: 8015 HDFS Write: 2511 SUCCESS INFO : Total MapReduce CPU Time Spent: 4 seconds 300 msec INFO : Completed executing command(queryId=hive_20200519100412_12f2b39c-45f4-4f4c-9261-32ca86fa28db); Time taken: 297.35 seconds 2020-05-19 10:11:54,635 DEBUG transport.TSaslTransport (TSaslTransport.java:flush(496)) - writing data length: 117 2020-05-19 10:11:54,638 DEBUG transport.TSaslTransport (TSaslTransport.java:readFrame(457)) - CLIENT: reading data length: 96 2020-05-19 10:11:54,649 DEBUG transport.TSaslTransport (TSaslTransport.java:flush(496)) - writing data length: 96 2020-05-19 10:11:54,714 DEBUG transport.TSaslTransport (TSaslTransport.java:readFrame(457)) - CLIENT: reading data length: 42 Error: Error while processing statement: FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.MoveTask. Exception when loading 2 in table dzdz_fpxx_dzfp with loadPath=hdfs://mycluster/warehouse/tablespace/managed/hive/dzdz_fpxx_dzfp/.hive-staging_hive_2020-05-19_10-04-12_468_7695595367555279265-3/-ext-10000 (state=08S01,code=1)
I find error log as follows in hive metastore:
2020-05-19 08:37:23,957 ERROR metastore.RetryingHMSHandler (RetryingHMSHandler.java:invokeInternal(175)) - NoSuchObjectException(message:partition values=[ujm]) at org.apache.hadoop.hive.metastore.ObjectStore.getPartition(ObjectStore.java:2003) at sun.reflect.GeneratedMethodAccessor15.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.hive.metastore.RawStoreProxy.invoke(RawStoreProxy.java:101) at com.sun.proxy.$Proxy26.getPartition(Unknown Source) at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.get_partition(HiveMetaStore.java:3553) at org.apache.hadoop.hive.metastore.events.InsertEvent.<init>(InsertEvent.java:62) at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.fire_listener_event(HiveMetaStore.java:6737) at sun.reflect.GeneratedMethodAccessor14.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:140) at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:99) at com.sun.proxy.$Proxy28.fire_listener_event(Unknown Source) at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor$fire_listener_event.getResult(ThriftHiveMetastore.java:14208) at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor$fire_listener_event.getResult(ThriftHiveMetastore.java:14193) at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39) at org.apache.hadoop.hive.metastore.TUGIBasedProcessor$1.run(TUGIBasedProcessor.java:110) at org.apache.hadoop.hive.metastore.TUGIBasedProcessor$1.run(TUGIBasedProcessor.java:106) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875) at org.apache.hadoop.hive.metastore.TUGIBasedProcessor.process(TUGIBasedProcessor.java:118) at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 2020-05-19 08:37:23,957 DEBUG log.PerfLogger (PerfLogger.java:PerfLogEnd(179)) - </PERFLOG method=fire_listener_event start=1589848643184 end=1589848643957 duration=773 from=org.apache.hadoop.hive.metastore.RetryingHMSHandler threadId=56 retryCount=-1 error=true> 2020-05-19 08:37:23,958 ERROR thrift.ProcessFunction (ProcessFunction.java:process(41)) - Internal error processing fire_listener_event org.apache.hadoop.hive.metastore.api.NoSuchObjectException: partition values=[ujm] at org.apache.hadoop.hive.metastore.ObjectStore.getPartition(ObjectStore.java:2003) ~[hive-exec-2.1.1-cdh6.2.0.jar:2.1.1-cdh6.2.0] at sun.reflect.GeneratedMethodAccessor15.invoke(Unknown Source) ~[?:?] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0-internal] at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0-internal] at org.apache.hadoop.hive.metastore.RawStoreProxy.invoke(RawStoreProxy.java:101) ~[hive-exec-2.1.1-cdh6.2.0.jar:2.1.1-cdh6.2.0] at com.sun.proxy.$Proxy26.getPartition(Unknown Source) ~[?:?] at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.get_partition(HiveMetaStore.java:3553) ~[hive-exec-2.1.1-cdh6.2.0.jar:2.1.1-cdh6.2.0] at org.apache.hadoop.hive.metastore.events.InsertEvent.<init>(InsertEvent.java:62) ~[hive-exec-2.1.1-cdh6.2.0.jar:2.1.1-cdh6.2.0] at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.fire_listener_event(HiveMetaStore.java:6737) ~[hive-exec-2.1.1-cdh6.2.0.jar:2.1.1-cdh6.2.0] at sun.reflect.GeneratedMethodAccessor14.invoke(Unknown Source) ~[?:?] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0-internal] at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0-internal] at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invokeInternal(RetryingHMSHandler.java:140) ~[hive-exec-2.1.1-cdh6.2.0.jar:2.1.1-cdh6.2.0] at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:99) ~[hive-exec-2.1.1-cdh6.2.0.jar:2.1.1-cdh6.2.0] at com.sun.proxy.$Proxy28.fire_listener_event(Unknown Source) ~[?:?] at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor$fire_listener_event.getResult(ThriftHiveMetastore.java:14208) ~[hive-exec-2.1.1-cdh6.2.0.jar:2.1.1-cdh6.2.0] at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor$fire_listener_event.getResult(ThriftHiveMetastore.java:14193) ~[hive-exec-2.1.1-cdh6.2.0.jar:2.1.1-cdh6.2.0] at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39) [hive-exec-2.1.1-cdh6.2.0.jar:2.1.1-cdh6.2.0] at org.apache.hadoop.hive.metastore.TUGIBasedProcessor$1.run(TUGIBasedProcessor.java:110) [hive-exec-2.1.1-cdh6.2.0.jar:2.1.1-cdh6.2.0] at org.apache.hadoop.hive.metastore.TUGIBasedProcessor$1.run(TUGIBasedProcessor.java:106) [hive-exec-2.1.1-cdh6.2.0.jar:2.1.1-cdh6.2.0] at java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0-internal] at javax.security.auth.Subject.doAs(Subject.java:422) [?:1.8.0-internal] at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1875) [hadoop-common-3.0.0-cdh6.2.0.jar:?] at org.apache.hadoop.hive.metastore.TUGIBasedProcessor.process(TUGIBasedProcessor.java:118) [hive-exec-2.1.1-cdh6.2.0.jar:2.1.1-cdh6.2.0] at org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:286) [hive-exec-2.1.1-cdh6.2.0.jar:2.1.1-cdh6.2.0] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0-internal] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0-internal] at java.lang.Thread.run(Thread.java:748) [?:1.8.0-internal] 2020-05-19 08:37:23,976 DEBUG ipc.Client (ClientCache.java:stopClient(97)) - stopping client from cache: org.apache.hadoop.ipc.Client@2c66e678
I guess metastore could not find partition,and disconnect connection.
Attachments
Issue Links
- is fixed by
-
HIVE-15642 Replicate Insert Overwrites, Dynamic Partition Inserts and Loads
- Closed