Details
-
Bug
-
Status: Closed
-
Minor
-
Resolution: Won't Fix
-
2.1.0
-
None
-
None
-
Spark 2.3.2
Description
Create table , load data and perform continous update operation on the table.
0: jdbc:hive2://10.20.255.171:23040> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 int) stored as carbondata TBLPROPERTIES ("TABLE_BLOCKSIZE"= "256 MB",'flat_folder'='true');
---------+
Result |
---------+
---------+
No rows selected (0.177 seconds)
0: jdbc:hive2://10.20.255.171:23040> LOAD DATA inpath 'hdfs://hacluster/chetan/2000_UniqData.csv' INTO table uniqdata options('DELIMITER'=',', 'FILEHEADER'='CUST_ID, CUST_NAME, ACTIVE_EMUI_VERSION, DOB, DOJ, BIGINT_COLUMN1, BIGINT_COLUMN2, DECIMAL_COLUMN1, DECIMAL_COLUMN2, Double_COLUMN1, Double_COLUMN2, INTEGER_COLUMN1','TIMESTAMPFORMAT'='yyyy-MM-dd HH:mm:ss','BAD_RECORDS_ACTION'='FORCE');
---------+
Result |
---------+
---------+
No rows selected (1.484 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1<123372036856;
--------------------+
Updated Row Count |
--------------------+
2 |
--------------------+
1 row selected (3.294 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1>123372038852;
--------------------+
Updated Row Count |
--------------------+
1 |
--------------------+
1 row selected (3.467 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1<=123372036859;
--------------------+
Updated Row Count |
--------------------+
9 |
--------------------+
1 row selected (3.349 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1>=123372038846;
--------------------+
Updated Row Count |
--------------------+
8 |
--------------------+
1 row selected (3.259 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1 == 123372038845;
--------------------+
Updated Row Count |
--------------------+
1 |
--------------------+
1 row selected (4.164 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1 like '123%';
--------------------+
Updated Row Count |
--------------------+
2000 |
--------------------+
1 row selected (3.695 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1 between 123372038849 AND 123372038855;
--------------------+
Updated Row Count |
--------------------+
5 |
--------------------+
1 row selected (3.228 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1 = 123372038845 OR false;
--------------------+
Updated Row Count |
--------------------+
1 |
--------------------+
1 row selected (3.548 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1 = 123372038849 AND true;
--------------------+
Updated Row Count |
--------------------+
1 |
--------------------+
1 row selected (3.321 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1 not between (123372038849) AND (12337203885);
--------------------+
Updated Row Count |
--------------------+
4025 |
--------------------+
1 row selected (3.718 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name<'CUST_NAME_01990';
--------------------+
Updated Row Count |
--------------------+
5978 |
--------------------+
1 row selected (4.109 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name>'CUST_NAME_01990';
--------------------+
Updated Row Count |
--------------------+
6022 |
--------------------+
1 row selected (3.643 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name<='CUST_NAME_01990';
--------------------+
Updated Row Count |
--------------------+
5981 |
--------------------+
1 row selected (3.713 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name >='CUST_NAME_01990';
--------------------+
Updated Row Count |
--------------------+
12050 |
--------------------+
1 row selected (4.019 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name=='CUST_NAME_01990';
--------------------+
Updated Row Count |
--------------------+
3 |
--------------------+
1 row selected (3.327 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name like 'CUST_NAME_0199%';
--------------------+
Updated Row Count |
--------------------+
47 |
--------------------+
1 row selected (3.47 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name between 'CUST_NAME_01990' AND 'CUST_NAME_01995';
--------------------+
Updated Row Count |
--------------------+
26 |
--------------------+
1 row selected (3.512 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name = 'CUST_NAME_01990' OR false;
--------------------+
Updated Row Count |
--------------------+
3 |
--------------------+
1 row selected (3.589 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name='cust_00033' AND false;
--------------------+
Updated Row Count |
--------------------+
0 |
--------------------+
1 row selected (0.462 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name not between 'CUST_NAME_01990' AND 'CUST_NAME_1999';
--------------------+
Updated Row Count |
--------------------+
18107 |
--------------------+
1 row selected (4.353 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100.00) where decimal_column1<12345678917.1234000000;
--------------------+
Updated Row Count |
--------------------+
224 |
--------------------+
1 row selected (3.494 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100.00) where decimal_column1>12345.6808982656;
--------------------+
Updated Row Count |
--------------------+
23964 |
--------------------+
1 row selected (4.677 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100.00) where decimal_column1<=12345678917.1234000000;
--------------------+
Updated Row Count |
--------------------+
24247 |
--------------------+
1 row selected (4.717 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100.00) where decimal_column1>=12345.6807932656;
--------------------+
Updated Row Count |
--------------------+
6025 |
--------------------+
1 row selected (3.754 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100.00) where decimal_column1=12345.6807332656;
--------------------+
Updated Row Count |
--------------------+
0 |
--------------------+
1 row selected (2.198 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100) where decimal_column1 like '123%';
--------------------+
Updated Row Count |
--------------------+
6025 |
--------------------+
1 row selected (4.173 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100) where decimal_column1 between 12345.6808882656 AND 12345.6809002656;
--------------------+
Updated Row Count |
--------------------+
0 |
--------------------+
1 row selected (2.113 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100) where (decimal_column1) = (12345678920.1234000000) OR false;
--------------------+
Updated Row Count |
--------------------+
3 |
--------------------+
1 row selected (3.798 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100) where decimal_column1 = 12345.6808782656 OR false;
--------------------+
Updated Row Count |
--------------------+
0 |
--------------------+
1 row selected (2.126 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100.0) where decimal_column1 not between 12345.6808882656 AND 12345.6809002656;
--------------------+
Updated Row Count |
--------------------+
42325 |
--------------------+
1 row selected (5.33 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(10000) where cust_id<10987;
--------------------+
Updated Row Count |
--------------------+
47870 |
--------------------+
1 row selected (5.907 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(1000) where cust_id>10987;
--------------------+
Updated Row Count |
--------------------+
456 |
--------------------+
1 row selected (3.596 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(100) where cust_id<=10987;
--------------------+
Updated Row Count |
--------------------+
96220 |
--------------------+
1 row selected (7.291 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(100) where cust_id>=10987;
--------------------+
Updated Row Count |
--------------------+
480 |
--------------------+
1 row selected (3.597 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(100) where cust_id==10987;
--------------------+
Updated Row Count |
--------------------+
24 |
--------------------+
1 row selected (3.534 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(100) where cust_id like '109%';
--------------------+
Updated Row Count |
--------------------+
2568 |
--------------------+
1 row selected (3.909 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(100) where cust_id between 10987 AND 10999;
--------------------+
Updated Row Count |
--------------------+
480 |
--------------------+
1 row selected (3.717 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(100) where cust_id = 10987 OR false;
--------------------+
Updated Row Count |
--------------------+
24 |
--------------------+
1 row selected (3.535 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(100) where cust_id = 10987 AND false;
--------------------+
Updated Row Count |
--------------------+
0 |
--------------------+
1 row selected (0.467 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(100) where cust_id not between 10987 AND 10999;
--------------------+
Updated Row Count |
--------------------+
147666 |
--------------------+
1 row selected (9.295 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(ASCII(cust_name)) where cust_id=10903;
--------------------+
Updated Row Count |
--------------------+
24 |
--------------------+
1 row selected (3.562 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(CONCAT(cust_name,ACTIVE_EMUI_VERSION)) where cust_id=10903;
--------------------+
Updated Row Count |
--------------------+
48 |
--------------------+
1 row selected (3.785 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(CONCAT('Mr.', cust_name)) where cust_id=10903;
--------------------+
Updated Row Count |
--------------------+
72 |
--------------------+
1 row selected (3.509 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(concat_ws('/',cust_name)) where cust_id=10902;
--------------------+
Updated Row Count |
--------------------+
24 |
--------------------+
1 row selected (3.525 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(concat_ws('/',cust_name,active_emui_version)) where cust_id=10902;
--------------------+
Updated Row Count |
--------------------+
48 |
--------------------+
1 row selected (3.537 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(LENGTH(cust_name)) where cust_id=10903;
--------------------+
Updated Row Count |
--------------------+
96 |
--------------------+
1 row selected (3.627 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(LENGTH(bigint_column1)) where cust_id=10903;
--------------------+
Updated Row Count |
--------------------+
120 |
--------------------+
1 row selected (3.61 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(LTRIM(cust_name)) where cust_id=10903;
--------------------+
Updated Row Count |
--------------------+
144 |
--------------------+
1 row selected (3.705 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(FORMAT_NUMBER(bigint_column1, 2)) where cust_id=10903;
--------------------+
Updated Row Count |
--------------------+
168 |
--------------------+
1 row selected (3.63 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(REVERSE(cust_name)) where cust_id=10903;
--------------------+
Updated Row Count |
--------------------+
192 |
--------------------+
1 row selected (3.655 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(REPEAT(cust_name,2)) where cust_id=10903;
--------------------+
Updated Row Count |
--------------------+
216 |
--------------------+
1 row selected (3.677 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('A J Styles') where cust_id=10902;
--------------------+
Updated Row Count |
--------------------+
72 |
--------------------+
1 row selected (3.674 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(cust_id) where cust_id=10902;
--------------------+
Updated Row Count |
--------------------+
96 |
--------------------+
1 row selected (3.643 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=('deepti') where cust_id=10903;
--------------------+
Updated Row Count |
--------------------+
240 |
--------------------+
1 row selected (3.756 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(21474836479999999) where cust_name='CUST_NAME_01999';
--------------------+
Updated Row Count |
--------------------+
150 |
--------------------+
1 row selected (3.727 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(-214748364867) where cust_name='CUST_NAME_01998';
--------------------+
Updated Row Count |
--------------------+
125 |
--------------------+
1 row selected (3.741 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(-9223372036854770000) where cust_name='CUST_NAME_01998';
--------------------+
Updated Row Count |
--------------------+
250 |
--------------------+
1 row selected (3.696 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column2)=(922337203685479999999999) where cust_name='CUST_NAME_01998';
--------------------+
Updated Row Count |
--------------------+
375 |
--------------------+
1 row selected (3.688 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (double_column1)=(4.9E-32) where cust_name='CUST_NAME_01998';
--------------------+
Updated Row Count |
--------------------+
500 |
--------------------+
1 row selected (3.889 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(1234567890123456789012.123456789012345) where cust_name='CUST_NAME_01998';
--------------------+
Updated Row Count |
--------------------+
625 |
--------------------+
1 row selected (11.982 seconds)
0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (dob)=('2016-13-31 12:00:00') where cust_name='CUST_NAME_01998';
--------------------+
Updated Row Count |
--------------------+
750 |
--------------------+
1 row selected (5.614 seconds)
After some time the update fails.
0: jdbc:hive2://10.20.255.171:23040> Update uniqdata set (CUST_ID) = ((select ASCII(CUST_NAME) from uniqdata where cust_name='CUST_NAME_00060'));
Error: java.lang.RuntimeException: Update operation failed. more than one row returned by a subquery used as an expression:
Subquery subquery980
+- *(1) Project ascii(CUST_NAME#45) AS ascii(CUST_NAME)#982
+- *(1) Filter (isnotnull(cust_name#45) && (cust_name#45 = CUST_NAME_00060))
+- *(1) FileScan carbondata chetan.uniqdatacust_name#45 PushedFilters: [IsNotNull(cust_name), EqualTo(cust_name,CUST_NAME_00060)], ReadSchema: struct<CUST_NAME:string> (state=,code=0)
Log-
2020-11-07 00:30:09,167 | ERROR | [HiveServer2-Background-Pool: Thread-363] | Exception in update operation | org.apache.spark.sql.execution.command.mutation.CarbonProjectForUpdateCommand.processData(CarbonProjectForUpdateCommand.scala:228)2020-11-07 00:30:09,167 | ERROR | [HiveServer2-Background-Pool: Thread-363] | Exception in update operation | org.apache.spark.sql.execution.command.mutation.CarbonProjectForUpdateCommand.processData(CarbonProjectForUpdateCommand.scala:228)java.lang.RuntimeException: more than one row returned by a subquery used as an expression:Subquery subquery980+- *(1) Project ascii(CUST_NAME#45) AS ascii(CUST_NAME)#982 +- *(1) Filter (isnotnull(cust_name#45) && (cust_name#45 = CUST_NAME_00060)) +- *(1) FileScan carbondata chetan.uniqdatacust_name#45 PushedFilters: [IsNotNull(cust_name), EqualTo(cust_name,CUST_NAME_00060)], ReadSchema: struct<CUST_NAME:string>
at scala.sys.package$.error(package.scala:27) at org.apache.spark.sql.execution.ScalarSubquery.updateResult(subquery.scala:69) at org.apache.spark.sql.execution.SparkPlan$$anonfun$waitForSubqueries$1.apply(SparkPlan.scala:185) at org.apache.spark.sql.execution.SparkPlan$$anonfun$waitForSubqueries$1.apply(SparkPlan.scala:184) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) at org.apache.spark.sql.execution.SparkPlan.waitForSubqueries(SparkPlan.scala:184) at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:154) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152) at org.apache.spark.sql.execution.CodegenSupport$class.produce(WholeStageCodegenExec.scala:83) at org.apache.spark.sql.execution.ProjectExec.produce(basicPhysicalOperators.scala:35) at org.apache.spark.sql.execution.WholeStageCodegenExec.doCodeGen(WholeStageCodegenExec.scala:524) at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:576) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:131) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:127) at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152) at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:127) at org.apache.spark.sql.execution.columnar.InMemoryRelation.buildBuffers(InMemoryRelation.scala:107) at org.apache.spark.sql.execution.columnar.InMemoryRelation.<init>(InMemoryRelation.scala:102) at org.apache.spark.sql.execution.columnar.InMemoryRelation$.apply(InMemoryRelation.scala:43) at org.apache.spark.sql.execution.CacheManager$$anonfun$cacheQuery$1.apply(CacheManager.scala:97) at org.apache.spark.sql.execution.CacheManager.writeLock(CacheManager.scala:67) at org.apache.spark.sql.execution.CacheManager.cacheQuery(CacheManager.scala:91) at org.apache.spark.sql.Dataset.persist(Dataset.scala:2929) at org.apache.spark.sql.execution.command.mutation.CarbonProjectForUpdateCommand.processData(CarbonProjectForUpdateCommand.scala:140) at org.apache.spark.sql.execution.command.DataCommand$$anonfun$run$2.apply(package.scala:132) at org.apache.spark.sql.execution.command.DataCommand$$anonfun$run$2.apply(package.scala:132) at org.apache.spark.sql.execution.command.Auditable$class.runWithAudit(package.scala:104) at org.apache.spark.sql.execution.command.DataCommand.runWithAudit(package.scala:130) at org.apache.spark.sql.execution.command.DataCommand.run(package.scala:132) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68) at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:79) at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:190) at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:190) at org.apache.spark.sql.Dataset$$anonfun$52.apply(Dataset.scala:3259) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:77) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3258) at org.apache.spark.sql.Dataset.<init>(Dataset.scala:190) at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:75) at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642) at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:694) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:232) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:175) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:171) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:185) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745)2020-11-07 00:30:09,173 | INFO | [HiveServer2-Background-Pool: Thread-363] | updateLock unlocked successfully after update uniqdata | org.apache.spark.sql.execution.command.mutation.CarbonProjectForUpdateCommand.processData(CarbonProjectForUpdateCommand.scala:248)2020-11-07 00:30:09,178 | INFO | [HiveServer2-Background-Pool: Thread-363] | compactionLock unlocked successfully after update uniqdata | org.apache.spark.sql.execution.command.mutation.CarbonProjectForUpdateCommand.processData(CarbonProjectForUpdateCommand.scala:254)2020-11-07 00:30:09,182 | INFO | [HiveServer2-Background-Pool: Thread-363] | Metadata lock has been successfully released | org.apache.carbondata.core.locks.CarbonLockUtil.fileUnlock(CarbonLockUtil.java:49)2020-11-07 00:30:09,189 | AUDIT | [HiveServer2-Background-Pool: Thread-363] | {"time":"November 7, 2020 12:30:09 AM CST","username":"anonymous","opName":"UPDATE DATA","opId":"2805904551130888","opStatus":"FAILED","opTime":"1313 ms","table":"chetan.uniqdata","extraInfo":{"Exception":"java.lang.RuntimeException","Message":"Update operation failed. more than one row returned by a subquery used as an expression:\nSubquery subquery980\n+- *(1) Project ascii(CUST_NAME#45) AS ascii(CUST_NAME)#982\n - *(1) Filter (isnotnull(cust_name#45) && (cust_name#45 = CUST_NAME_00060))\n +- *(1) FileScan carbondata chetan.uniqdatacust_name#45 PushedFilters: [IsNotNull(cust_name), EqualTo(cust_name,CUST_NAME_00060)], ReadSchema: struct<CUST_NAME:string>\n"}} | org.apache.carbondata.processing.util.Auditor.logOperationEnd(Auditor.java:97)2020-11-07 00:30:09,191 | ERROR | [HiveServer2-Background-Pool: Thread-363] | Error executing query, currentState RUNNING, | org.apache.spark.internal.Logging$class.logError(Logging.scala:91)java.lang.RuntimeException: Update operation failed. more than one row returned by a subquery used as an expression:Subquery subquery980- *(1) Project ascii(CUST_NAME#45) AS ascii(CUST_NAME)#982 +- *(1) Filter (isnotnull(cust_name#45) && (cust_name#45 = CUST_NAME_00060)) +- *(1) FileScan carbondata chetan.uniqdatacust_name#45 PushedFilters: [IsNotNull(cust_name), EqualTo(cust_name,CUST_NAME_00060)], ReadSchema: struct<CUST_NAME:string>
at scala.sys.package$.error(package.scala:27) at org.apache.spark.sql.execution.command.mutation.CarbonProjectForUpdateCommand.processData(CarbonProjectForUpdateCommand.scala:232) at org.apache.spark.sql.execution.command.DataCommand$$anonfun$run$2.apply(package.scala:132) at org.apache.spark.sql.execution.command.DataCommand$$anonfun$run$2.apply(package.scala:132) at org.apache.spark.sql.execution.command.Auditable$class.runWithAudit(package.scala:104) at org.apache.spark.sql.execution.command.DataCommand.runWithAudit(package.scala:130) at org.apache.spark.sql.execution.command.DataCommand.run(package.scala:132) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68) at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:79) at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:190) at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:190) at org.apache.spark.sql.Dataset$$anonfun$52.apply(Dataset.scala:3259) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:77) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3258) at org.apache.spark.sql.Dataset.<init>(Dataset.scala:190) at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:75) at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642) at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:694) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:232) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:175) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:171) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:185) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745)2020-11-07 00:30:09,192 | ERROR | [HiveServer2-Background-Pool: Thread-363] | Error running hive query: | org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:179)org.apache.hive.service.cli.HiveSQLException: java.lang.RuntimeException: Update operation failed. more than one row returned by a subquery used as an expression:Subquery subquery980+- *(1) Project ascii(CUST_NAME#45) AS ascii(CUST_NAME)#982 +- *(1) Filter (isnotnull(cust_name#45) && (cust_name#45 = CUST_NAME_00060)) +- *(1) FileScan carbondata chetan.uniqdatacust_name#45 PushedFilters: [IsNotNull(cust_name), EqualTo(cust_name,CUST_NAME_00060)], ReadSchema: struct<CUST_NAME:string>
at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:269) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:175) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:171) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:185) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745)2020-11-07 00:30:09,195 | INFO | [HiveServer2-Handler-Pool: Thread-83] | Asked to cancel job group 60a4f869-1300-4a8a-a575-196f32184ed1 | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54)