Chetan Bhat created CARBONDATA-4048:
--------------------------------------- Summary: Update fails after continous update operations with error Key: CARBONDATA-4048 URL: https://issues.apache.org/jira/browse/CARBONDATA-4048 Project: CarbonData Issue Type: Bug Components: data-query Affects Versions: 2.1.0 Environment: Spark 2.3.2 Reporter: Chetan Bhat Create table , load data and perform continous update operation on the table. 0: jdbc:hive2://10.20.255.171:23040> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 int) stored as carbondata TBLPROPERTIES ("TABLE_BLOCKSIZE"= "256 MB",'flat_folder'='true'); +---------+--+ | Result | +---------+--+ +---------+--+ No rows selected (0.177 seconds) 0: jdbc:hive2://10.20.255.171:23040> LOAD DATA inpath 'hdfs://hacluster/chetan/2000_UniqData.csv' INTO table uniqdata options('DELIMITER'=',', 'FILEHEADER'='CUST_ID, CUST_NAME, ACTIVE_EMUI_VERSION, DOB, DOJ, BIGINT_COLUMN1, BIGINT_COLUMN2, DECIMAL_COLUMN1, DECIMAL_COLUMN2, Double_COLUMN1, Double_COLUMN2, INTEGER_COLUMN1','TIMESTAMPFORMAT'='yyyy-MM-dd HH:mm:ss','BAD_RECORDS_ACTION'='FORCE'); +---------+--+ | Result | +---------+--+ +---------+--+ No rows selected (1.484 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1<123372036856; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 2 | +--------------------+--+ 1 row selected (3.294 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1>123372038852; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 1 | +--------------------+--+ 1 row selected (3.467 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1<=123372036859; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 9 | +--------------------+--+ 1 row selected (3.349 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1>=123372038846; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 8 | +--------------------+--+ 1 row selected (3.259 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1 == 123372038845; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 1 | +--------------------+--+ 1 row selected (4.164 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1 like '123%'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 2000 | +--------------------+--+ 1 row selected (3.695 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1 between 123372038849 AND 123372038855; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 5 | +--------------------+--+ 1 row selected (3.228 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1 = 123372038845 OR false; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 1 | +--------------------+--+ 1 row selected (3.548 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1 = 123372038849 AND true; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 1 | +--------------------+--+ 1 row selected (3.321 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(100) where bigint_column1 not between (123372038849) AND (12337203885); +--------------------+--+ | Updated Row Count | +--------------------+--+ | 4025 | +--------------------+--+ 1 row selected (3.718 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name<'CUST_NAME_01990'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 5978 | +--------------------+--+ 1 row selected (4.109 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name>'CUST_NAME_01990'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 6022 | +--------------------+--+ 1 row selected (3.643 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name<='CUST_NAME_01990'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 5981 | +--------------------+--+ 1 row selected (3.713 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name >='CUST_NAME_01990'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 12050 | +--------------------+--+ 1 row selected (4.019 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name=='CUST_NAME_01990'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 3 | +--------------------+--+ 1 row selected (3.327 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name like 'CUST_NAME_0199%'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 47 | +--------------------+--+ 1 row selected (3.47 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name between 'CUST_NAME_01990' AND 'CUST_NAME_01995'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 26 | +--------------------+--+ 1 row selected (3.512 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name = 'CUST_NAME_01990' OR false; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 3 | +--------------------+--+ 1 row selected (3.589 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name='cust_00033' AND false; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 0 | +--------------------+--+ 1 row selected (0.462 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('deepti') where cust_name not between 'CUST_NAME_01990' AND 'CUST_NAME_1999'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 18107 | +--------------------+--+ 1 row selected (4.353 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100.00) where decimal_column1<12345678917.1234000000; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 224 | +--------------------+--+ 1 row selected (3.494 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100.00) where decimal_column1>12345.6808982656; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 23964 | +--------------------+--+ 1 row selected (4.677 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100.00) where decimal_column1<=12345678917.1234000000; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 24247 | +--------------------+--+ 1 row selected (4.717 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100.00) where decimal_column1>=12345.6807932656; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 6025 | +--------------------+--+ 1 row selected (3.754 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100.00) where decimal_column1=12345.6807332656; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 0 | +--------------------+--+ 1 row selected (2.198 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100) where decimal_column1 like '123%'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 6025 | +--------------------+--+ 1 row selected (4.173 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100) where decimal_column1 between 12345.6808882656 AND 12345.6809002656; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 0 | +--------------------+--+ 1 row selected (2.113 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100) where (decimal_column1) = (12345678920.1234000000) OR false; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 3 | +--------------------+--+ 1 row selected (3.798 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100) where decimal_column1 = 12345.6808782656 OR false; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 0 | +--------------------+--+ 1 row selected (2.126 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(100.0) where decimal_column1 not between 12345.6808882656 AND 12345.6809002656; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 42325 | +--------------------+--+ 1 row selected (5.33 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(10000) where cust_id<10987; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 47870 | +--------------------+--+ 1 row selected (5.907 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(1000) where cust_id>10987; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 456 | +--------------------+--+ 1 row selected (3.596 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(100) where cust_id<=10987; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 96220 | +--------------------+--+ 1 row selected (7.291 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(100) where cust_id>=10987; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 480 | +--------------------+--+ 1 row selected (3.597 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(100) where cust_id==10987; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 24 | +--------------------+--+ 1 row selected (3.534 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(100) where cust_id like '109%'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 2568 | +--------------------+--+ 1 row selected (3.909 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(100) where cust_id between 10987 AND 10999; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 480 | +--------------------+--+ 1 row selected (3.717 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(100) where cust_id = 10987 OR false; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 24 | +--------------------+--+ 1 row selected (3.535 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(100) where cust_id = 10987 AND false; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 0 | +--------------------+--+ 1 row selected (0.467 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(100) where cust_id not between 10987 AND 10999; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 147666 | +--------------------+--+ 1 row selected (9.295 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(ASCII(cust_name)) where cust_id=10903; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 24 | +--------------------+--+ 1 row selected (3.562 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(CONCAT(cust_name,ACTIVE_EMUI_VERSION)) where cust_id=10903; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 48 | +--------------------+--+ 1 row selected (3.785 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(CONCAT('Mr.', cust_name)) where cust_id=10903; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 72 | +--------------------+--+ 1 row selected (3.509 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(concat_ws('/',cust_name)) where cust_id=10902; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 24 | +--------------------+--+ 1 row selected (3.525 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(concat_ws('/',cust_name,active_emui_version)) where cust_id=10902; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 48 | +--------------------+--+ 1 row selected (3.537 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(LENGTH(cust_name)) where cust_id=10903; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 96 | +--------------------+--+ 1 row selected (3.627 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(LENGTH(bigint_column1)) where cust_id=10903; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 120 | +--------------------+--+ 1 row selected (3.61 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(LTRIM(cust_name)) where cust_id=10903; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 144 | +--------------------+--+ 1 row selected (3.705 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(FORMAT_NUMBER(bigint_column1, 2)) where cust_id=10903; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 168 | +--------------------+--+ 1 row selected (3.63 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(REVERSE(cust_name)) where cust_id=10903; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 192 | +--------------------+--+ 1 row selected (3.655 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(REPEAT(cust_name,2)) where cust_id=10903; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 216 | +--------------------+--+ 1 row selected (3.677 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=('A J Styles') where cust_id=10902; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 72 | +--------------------+--+ 1 row selected (3.674 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_name)=(cust_id) where cust_id=10902; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 96 | +--------------------+--+ 1 row selected (3.643 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=('deepti') where cust_id=10903; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 240 | +--------------------+--+ 1 row selected (3.756 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(21474836479999999) where cust_name='CUST_NAME_01999'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 150 | +--------------------+--+ 1 row selected (3.727 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (cust_id)=(-214748364867) where cust_name='CUST_NAME_01998'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 125 | +--------------------+--+ 1 row selected (3.741 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column1)=(-9223372036854770000) where cust_name='CUST_NAME_01998'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 250 | +--------------------+--+ 1 row selected (3.696 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (bigint_column2)=(922337203685479999999999) where cust_name='CUST_NAME_01998'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 375 | +--------------------+--+ 1 row selected (3.688 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (double_column1)=(4.9E-32) where cust_name='CUST_NAME_01998'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 500 | +--------------------+--+ 1 row selected (3.889 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (decimal_column1)=(1234567890123456789012.123456789012345) where cust_name='CUST_NAME_01998'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 625 | +--------------------+--+ 1 row selected (11.982 seconds) 0: jdbc:hive2://10.20.255.171:23040> update uniqdata set (dob)=('2016-13-31 12:00:00') where cust_name='CUST_NAME_01998'; +--------------------+--+ | Updated Row Count | +--------------------+--+ | 750 | +--------------------+--+ 1 row selected (5.614 seconds) After some time the update fails. 0: jdbc:hive2://10.20.255.171:23040> Update uniqdata set (CUST_ID) = ((select ASCII(CUST_NAME) from uniqdata where cust_name='CUST_NAME_00060')); Error: java.lang.RuntimeException: Update operation failed. more than one row returned by a subquery used as an expression: Subquery subquery980 +- *(1) Project [ascii(CUST_NAME#45) AS ascii(CUST_NAME)#982] +- *(1) Filter (isnotnull(cust_name#45) && (cust_name#45 = CUST_NAME_00060)) +- *(1) FileScan carbondata chetan.uniqdata[cust_name#45] PushedFilters: [IsNotNull(cust_name), EqualTo(cust_name,CUST_NAME_00060)], ReadSchema: struct<CUST_NAME:string> (state=,code=0) Log- 2020-11-07 00:30:09,167 | ERROR | [HiveServer2-Background-Pool: Thread-363] | Exception in update operation | org.apache.spark.sql.execution.command.mutation.CarbonProjectForUpdateCommand.processData(CarbonProjectForUpdateCommand.scala:228)2020-11-07 00:30:09,167 | ERROR | [HiveServer2-Background-Pool: Thread-363] | Exception in update operation | org.apache.spark.sql.execution.command.mutation.CarbonProjectForUpdateCommand.processData(CarbonProjectForUpdateCommand.scala:228)java.lang.RuntimeException: more than one row returned by a subquery used as an expression:Subquery subquery980+- *(1) Project [ascii(CUST_NAME#45) AS ascii(CUST_NAME)#982] +- *(1) Filter (isnotnull(cust_name#45) && (cust_name#45 = CUST_NAME_00060)) +- *(1) FileScan carbondata chetan.uniqdata[cust_name#45] PushedFilters: [IsNotNull(cust_name), EqualTo(cust_name,CUST_NAME_00060)], ReadSchema: struct<CUST_NAME:string> at scala.sys.package$.error(package.scala:27) at org.apache.spark.sql.execution.ScalarSubquery.updateResult(subquery.scala:69) at org.apache.spark.sql.execution.SparkPlan$$anonfun$waitForSubqueries$1.apply(SparkPlan.scala:185) at org.apache.spark.sql.execution.SparkPlan$$anonfun$waitForSubqueries$1.apply(SparkPlan.scala:184) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) at org.apache.spark.sql.execution.SparkPlan.waitForSubqueries(SparkPlan.scala:184) at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:154) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152) at org.apache.spark.sql.execution.CodegenSupport$class.produce(WholeStageCodegenExec.scala:83) at org.apache.spark.sql.execution.ProjectExec.produce(basicPhysicalOperators.scala:35) at org.apache.spark.sql.execution.WholeStageCodegenExec.doCodeGen(WholeStageCodegenExec.scala:524) at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:576) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:131) at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:127) at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152) at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:127) at org.apache.spark.sql.execution.columnar.InMemoryRelation.buildBuffers(InMemoryRelation.scala:107) at org.apache.spark.sql.execution.columnar.InMemoryRelation.<init>(InMemoryRelation.scala:102) at org.apache.spark.sql.execution.columnar.InMemoryRelation$.apply(InMemoryRelation.scala:43) at org.apache.spark.sql.execution.CacheManager$$anonfun$cacheQuery$1.apply(CacheManager.scala:97) at org.apache.spark.sql.execution.CacheManager.writeLock(CacheManager.scala:67) at org.apache.spark.sql.execution.CacheManager.cacheQuery(CacheManager.scala:91) at org.apache.spark.sql.Dataset.persist(Dataset.scala:2929) at org.apache.spark.sql.execution.command.mutation.CarbonProjectForUpdateCommand.processData(CarbonProjectForUpdateCommand.scala:140) at org.apache.spark.sql.execution.command.DataCommand$$anonfun$run$2.apply(package.scala:132) at org.apache.spark.sql.execution.command.DataCommand$$anonfun$run$2.apply(package.scala:132) at org.apache.spark.sql.execution.command.Auditable$class.runWithAudit(package.scala:104) at org.apache.spark.sql.execution.command.DataCommand.runWithAudit(package.scala:130) at org.apache.spark.sql.execution.command.DataCommand.run(package.scala:132) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68) at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:79) at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:190) at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:190) at org.apache.spark.sql.Dataset$$anonfun$52.apply(Dataset.scala:3259) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:77) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3258) at org.apache.spark.sql.Dataset.<init>(Dataset.scala:190) at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:75) at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642) at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:694) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:232) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:175) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:171) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:185) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745)2020-11-07 00:30:09,173 | INFO | [HiveServer2-Background-Pool: Thread-363] | updateLock unlocked successfully after update uniqdata | org.apache.spark.sql.execution.command.mutation.CarbonProjectForUpdateCommand.processData(CarbonProjectForUpdateCommand.scala:248)2020-11-07 00:30:09,178 | INFO | [HiveServer2-Background-Pool: Thread-363] | compactionLock unlocked successfully after update uniqdata | org.apache.spark.sql.execution.command.mutation.CarbonProjectForUpdateCommand.processData(CarbonProjectForUpdateCommand.scala:254)2020-11-07 00:30:09,182 | INFO | [HiveServer2-Background-Pool: Thread-363] | Metadata lock has been successfully released | org.apache.carbondata.core.locks.CarbonLockUtil.fileUnlock(CarbonLockUtil.java:49)2020-11-07 00:30:09,189 | AUDIT | [HiveServer2-Background-Pool: Thread-363] | \{"time":"November 7, 2020 12:30:09 AM CST","username":"anonymous","opName":"UPDATE DATA","opId":"2805904551130888","opStatus":"FAILED","opTime":"1313 ms","table":"chetan.uniqdata","extraInfo":{"Exception":"java.lang.RuntimeException","Message":"Update operation failed. more than one row returned by a subquery used as an expression:\nSubquery subquery980\n+- *(1) Project [ascii(CUST_NAME#45) AS ascii(CUST_NAME)#982]\n +- *(1) Filter (isnotnull(cust_name#45) && (cust_name#45 = CUST_NAME_00060))\n +- *(1) FileScan carbondata chetan.uniqdata[cust_name#45] PushedFilters: [IsNotNull(cust_name), EqualTo(cust_name,CUST_NAME_00060)], ReadSchema: struct<CUST_NAME:string>\n"}} | org.apache.carbondata.processing.util.Auditor.logOperationEnd(Auditor.java:97)2020-11-07 00:30:09,191 | ERROR | [HiveServer2-Background-Pool: Thread-363] | Error executing query, currentState RUNNING, | org.apache.spark.internal.Logging$class.logError(Logging.scala:91)java.lang.RuntimeException: Update operation failed. more than one row returned by a subquery used as an expression:Subquery subquery980+- *(1) Project [ascii(CUST_NAME#45) AS ascii(CUST_NAME)#982] +- *(1) Filter (isnotnull(cust_name#45) && (cust_name#45 = CUST_NAME_00060)) +- *(1) FileScan carbondata chetan.uniqdata[cust_name#45] PushedFilters: [IsNotNull(cust_name), EqualTo(cust_name,CUST_NAME_00060)], ReadSchema: struct<CUST_NAME:string> at scala.sys.package$.error(package.scala:27) at org.apache.spark.sql.execution.command.mutation.CarbonProjectForUpdateCommand.processData(CarbonProjectForUpdateCommand.scala:232) at org.apache.spark.sql.execution.command.DataCommand$$anonfun$run$2.apply(package.scala:132) at org.apache.spark.sql.execution.command.DataCommand$$anonfun$run$2.apply(package.scala:132) at org.apache.spark.sql.execution.command.Auditable$class.runWithAudit(package.scala:104) at org.apache.spark.sql.execution.command.DataCommand.runWithAudit(package.scala:130) at org.apache.spark.sql.execution.command.DataCommand.run(package.scala:132) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68) at org.apache.spark.sql.execution.command.ExecutedCommandExec.executeCollect(commands.scala:79) at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:190) at org.apache.spark.sql.Dataset$$anonfun$6.apply(Dataset.scala:190) at org.apache.spark.sql.Dataset$$anonfun$52.apply(Dataset.scala:3259) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:77) at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3258) at org.apache.spark.sql.Dataset.<init>(Dataset.scala:190) at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:75) at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642) at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:694) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:232) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:175) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:171) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:185) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745)2020-11-07 00:30:09,192 | ERROR | [HiveServer2-Background-Pool: Thread-363] | Error running hive query: | org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:179)org.apache.hive.service.cli.HiveSQLException: java.lang.RuntimeException: Update operation failed. more than one row returned by a subquery used as an expression:Subquery subquery980+- *(1) Project [ascii(CUST_NAME#45) AS ascii(CUST_NAME)#982] +- *(1) Filter (isnotnull(cust_name#45) && (cust_name#45 = CUST_NAME_00060)) +- *(1) FileScan carbondata chetan.uniqdata[cust_name#45] PushedFilters: [IsNotNull(cust_name), EqualTo(cust_name,CUST_NAME_00060)], ReadSchema: struct<CUST_NAME:string> at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:269) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:175) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:171) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698) at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:185) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745)2020-11-07 00:30:09,195 | INFO | [HiveServer2-Handler-Pool: Thread-83] | Asked to cancel job group 60a4f869-1300-4a8a-a575-196f32184ed1 | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) -- This message was sent by Atlassian Jira (v8.3.4#803005) |
Free forum by Nabble | Edit this page |