[ https://issues.apache.org/jira/browse/CARBONDATA-1237?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] QiangCai reassigned CARBONDATA-1237: ------------------------------------ Assignee: QiangCai > UndeclaredThrowableException thrown when secondary index created on timestamp column for multiple loads(unique data) with enable.unsafe.columnpage = true > --------------------------------------------------------------------------------------------------------------------------------------------------------- > > Key: CARBONDATA-1237 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1237 > Project: CarbonData > Issue Type: Bug > Components: data-query > Affects Versions: 1.1.0 > Environment: SUSE 11 SP4 3 node cluster. > Reporter: Chetan Bhat > Assignee: QiangCai > > Steps: > Configure enable.unsafe.columnpage = true. > Create table having timestamp column. > Load data - 2 loads using different csv each csv having unique data. > Try to create secondary index on timestamp column. > 0: jdbc:hive2://10.19.91.225:22550/default> CREATE TABLE uniqdata (CUST_ID int,CUST_NAME String,ACTIVE_EMUI_VERSION string, DOB timestamp, DOJ timestamp, BIGINT_COLUMN1 bigint,BIGINT_COLUMN2 bigint,DECIMAL_COLUMN1 decimal(30,10), DECIMAL_COLUMN2 decimal(36,10),Double_COLUMN1 double, Double_COLUMN2 double,INTEGER_COLUMN1 int) STORED BY 'org.apache.carbondata.format' TBLPROPERTIES('table_blocksize'='128','include_dictionary'='CUST_NAME,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1,CUST_ID','sort_columns'='CUST_NAME'); > +---------+--+ > | Result | > +---------+--+ > +---------+--+ > No rows selected (1.017 seconds) > 0: jdbc:hive2://10.19.91.225:22550/default> LOAD DATA INPATH 'hdfs://hacluster/chetan/split1.csv' into table uniqdata OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1'); > +---------+--+ > | Result | > +---------+--+ > +---------+--+ > No rows selected (10.798 seconds) > 0: jdbc:hive2://10.19.91.225:22550/default> LOAD DATA INPATH 'hdfs://hacluster/chetan/split2.csv' into table uniqdata OPTIONS('DELIMITER'=',' , 'QUOTECHAR'='"','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='CUST_ID,CUST_NAME,ACTIVE_EMUI_VERSION,DOB,DOJ,BIGINT_COLUMN1,BIGINT_COLUMN2,DECIMAL_COLUMN1,DECIMAL_COLUMN2,Double_COLUMN1,Double_COLUMN2,INTEGER_COLUMN1'); > +---------+--+ > | Result | > +---------+--+ > +---------+--+ > No rows selected (16.006 seconds) > 0: jdbc:hive2://10.19.91.225:22550/default> CREATE INDEX indextable2 ON TABLE uniqdata (DOB) AS 'org.apache.carbondata.format'; > Actual Result(Issue ) - UndeclaredThrowableException thrown when secondary index tried to be created on timestamp column. > 0: jdbc:hive2://10.19.91.225:22550/default> CREATE INDEX indextable2 ON TABLE uniqdata (DOB) AS 'org.apache.carbondata.format'; > Error: java.lang.reflect.UndeclaredThrowableException (state=,code=0) > Driver Log - > 2017-06-27 14:06:34,557 | INFO | [pool-23-thread-83] | Parsing command: string | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) > 2017-06-27 14:06:34,557 | INFO | [pool-23-thread-83] | Parsing command: timestamp | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) > 2017-06-27 14:06:34,557 | INFO | [pool-23-thread-83] | Parsing command: timestamp | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) > 2017-06-27 14:06:34,557 | INFO | [pool-23-thread-83] | Parsing command: bigint | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) > 2017-06-27 14:06:34,558 | INFO | [pool-23-thread-83] | Parsing command: bigint | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) > 2017-06-27 14:06:34,558 | INFO | [pool-23-thread-83] | Parsing command: decimal(30,10) | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) > 2017-06-27 14:06:34,558 | INFO | [pool-23-thread-83] | Parsing command: decimal(36,10) | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) > 2017-06-27 14:06:34,558 | INFO | [pool-23-thread-83] | Parsing command: double | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) > 2017-06-27 14:06:34,559 | INFO | [pool-23-thread-83] | Parsing command: double | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) > 2017-06-27 14:06:34,559 | INFO | [pool-23-thread-83] | Parsing command: int | org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) > 2017-06-27 14:06:34,559 | INFO | [pool-23-thread-83] | Closed a connection to metastore, current connections: 0 | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.close(HiveMetaStoreClient.java:622) > 2017-06-27 14:06:34,560 | INFO | [pool-23-thread-83] | Trying to connect to metastore with URI thrift://10.19.91.224:21088 | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:469) > 2017-06-27 14:06:34,585 | INFO | [pool-23-thread-83] | Opened a connection to metastore, current connections: 1 | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:514) > 2017-06-27 14:06:34,585 | INFO | [pool-23-thread-83] | Connected to metastore. | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:593) > 2017-06-27 14:06:34,793 | AUDIT | [pool-23-thread-83] | [linux-225][chetan][Thread-776]Index creation with Database name [default] and Index name [indextable2] failed | org.apache.carbondata.common.logging.impl.StandardLogService.audit(StandardLogService.java:207) > 2017-06-27 14:06:34,793 | INFO | [pool-23-thread-83] | OperationId=75f45d62-5e90-41f8-8e61-186a5e8e58e3 Result=FAIL | org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:320) > 2017-06-27 14:06:34,793 | ERROR | [pool-23-thread-83] | Error executing query, currentState RUNNING, | org.apache.spark.internal.Logging$class.logError(Logging.scala:91) > java.lang.reflect.UndeclaredThrowableException > at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1787) > at org.apache.spark.util.Utils$.proxyOperate(Utils.scala:2688) > at org.apache.spark.util.ACLFileUtils$.changeOwnerRecursivelyAfterOperation(ACLFileUtils.scala:57) > at org.apache.spark.sql.execution.command.CreateIndexTable.run(carbonTableSchema.scala:359) > at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:59) > at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:57) > at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:75) > at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114) > at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114) > at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:135) > at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) > at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:132) > at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:113) > at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:125) > at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:125) > at org.apache.spark.sql.Dataset.<init>(Dataset.scala:185) > at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64) > at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:592) > at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:699) > at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:251) > at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:183) > at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:180) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1769) > at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:193) > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:748) > Caused by: java.util.concurrent.ExecutionException: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 98.0 failed 4 times, most recent failure: Lost task 0.3 in stage 98.0 (TID 199, linux-224, executor 30): ExecutorLostFailure (executor 30 exited caused by one of the running tasks) Reason: Container marked as failed: container_1497850829760_1804_01_000042 on host: linux-224. Exit status: 134. Diagnostics: Exception from container-launch. > Container id: container_1497850829760_1804_01_000042 > Exit code: 134 > Stack trace: ExitCodeException exitCode=134: > at org.apache.hadoop.util.Shell.runCommand(Shell.java:628) > at org.apache.hadoop.util.Shell.run(Shell.java:525) > at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:819) > at org.apache.hadoop.yarn.server.nodemanager.LinuxContainerExecutor.launchContainer(LinuxContainerExecutor.java:388) > at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:313) > at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:88) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:748) > Shell output: main : command provided 1 > main : run as user is spark2x > main : requested yarn user is spark2x > Before: Going to create dir > Container exited with a non-zero exit code 134 and the last 4096 bytes from the error logs are : > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in [jar:file:/opt/huawei/Bigdata/FusionInsight_Spark2x_V100R002C70/install/FusionInsight-Spark2x-2.1/spark/jars/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in [jar:file:/srv/BigData/hadoop/data1/nm/localdir/filecache/11/spark-archive-2x.zip/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in [jar:file:/opt/huawei/Bigdata/FusionInsight_HD_V100R002C70/install/FusionInsight-Hadoop-2.7.2/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] > Driver stacktrace: > at java.util.concurrent.FutureTask.report(FutureTask.java:122) > at java.util.concurrent.FutureTask.get(FutureTask.java:192) > at org.apache.carbondata.spark.rdd.SecondaryIndexCreator$$anonfun$createSecondaryIndex$1$$anonfun$apply$mcV$sp$1.apply(SecondaryIndexCreator.scala:106) > at org.apache.carbondata.spark.rdd.SecondaryIndexCreator$$anonfun$createSecondaryIndex$1$$anonfun$apply$mcV$sp$1.apply(SecondaryIndexCreator.scala:105) > at scala.collection.immutable.List.foreach(List.scala:381) > at org.apache.carbondata.spark.rdd.SecondaryIndexCreator$$anonfun$createSecondaryIndex$1.apply$mcV$sp(SecondaryIndexCreator.scala:105) > at scala.util.control.Breaks.breakable(Breaks.scala:38) > at org.apache.carbondata.spark.rdd.SecondaryIndexCreator$.createSecondaryIndex(SecondaryIndexCreator.scala:103) > at org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.secondaryIndexCallable(CarbonDataRDDFactory.scala:156) > at org.apache.carbondata.spark.rdd.CarbonDataRDDFactory$.createSecondaryIndex(CarbonDataRDDFactory.scala:103) > at org.apache.spark.sql.execution.command.loadDataForSecondaryIndex.run(carbonTableSchema.scala:491) > at org.apache.spark.sql.execution.command.CreateIndexTable$$anonfun$run$2.apply$mcV$sp(carbonTableSchema.scala:361) > at org.apache.spark.util.ACLFileUtils$$anonfun$changeOwnerRecursivelyAfterOperation$1.apply$mcV$sp(ACLFileUtils.scala:71) > at org.apache.spark.util.Utils$$anon$7.run(Utils.scala:2690) > at org.apache.spark.util.Utils$$anon$7.run(Utils.scala:2688) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1769) > ... 30 more > Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 98.0 failed 4 times, most recent failure: Lost task 0.3 in stage 98.0 (TID 199, linux-224, executor 30): ExecutorLostFailure (executor 30 exited caused by one of the running tasks) Reason: Container marked as failed: container_1497850829760_1804_01_000042 on host: linux-224. Exit status: 134. Diagnostics: Exception from container-launch. > Container id: container_1497850829760_1804_01_000042 > Exit code: 134 > Stack trace: ExitCodeException exitCode=134: > at org.apache.hadoop.util.Shell.runCommand(Shell.java:628) > at org.apache.hadoop.util.Shell.run(Shell.java:525) > at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:819) > at org.apache.hadoop.yarn.server.nodemanager.LinuxContainerExecutor.launchContainer(LinuxContainerExecutor.java:388) > at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:313) > at org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:88) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:748) > Shell output: main : command provided 1 > main : run as user is spark2x > main : requested yarn user is spark2x > Before: Going to create dir > Container exited with a non-zero exit code 134 and the last 4096 bytes from the error logs are : > SLF4J: Class path contains multiple SLF4J bindings. > SLF4J: Found binding in [jar:file:/opt/huawei/Bigdata/FusionInsight_Spark2x_V100R002C70/install/FusionInsight-Spark2x-2.1/spark/jars/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in [jar:file:/srv/BigData/hadoop/data1/nm/localdir/filecache/11/spark-archive-2x.zip/slf4j-log4j12-1.7.16.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: Found binding in [jar:file:/opt/huawei/Bigdata/FusionInsight_HD_V100R002C70/install/FusionInsight-Hadoop-2.7.2/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] > SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. > SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] > Driver stacktrace: > at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1478) > at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1466) > at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1465) > at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) > at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1465) > at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:813) > at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:813) > at scala.Option.foreach(Option.scala:257) > at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:813) > at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1693) > at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1648) > at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1637) > at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) > at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:639) > at org.apache.spark.SparkContext.runJob(SparkContext.scala:1946) > at org.apache.spark.SparkContext.runJob(SparkContext.scala:1959) > at org.apache.spark.SparkContext.runJob(SparkContext.scala:1972) > at org.apache.spark.SparkContext.runJob(SparkContext.scala:1986) > at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:937) > at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) > at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112) > at org.apache.spark.rdd.RDD.withScope(RDD.scala:363) > at org.apache.spark.rdd.RDD.collect(RDD.scala:936) > at org.apache.carbondata.spark.rdd.SecondaryIndexCreator$$anonfun$createSecondaryIndex$2$$anon$1.call(SecondaryIndexCreator.scala:92) > at org.apache.carbondata.spark.rdd.SecondaryIndexCreator$$anonfun$createSecondaryIndex$2$$anon$1.call(SecondaryIndexCreator.scala:77) > ... 4 more > 2017-06-27 14:06:34,794 | ERROR | [pool-23-thread-83] | Error running hive query: | org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:187) > org.apache.hive.service.cli.HiveSQLException: java.lang.reflect.UndeclaredThrowableException > at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:333) > at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:183) > at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1$$anon$2.run(SparkExecuteStatementOperation.scala:180) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:422) > at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1769) > at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$1.run(SparkExecuteStatementOperation.scala:193) > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:748) > 2017-06-27 14:06:36,996 | INFO | [main] | the jdbc URL switch open:jdbc:hive2://10.19.91.225:22550/default;principal=spark2x/[hidden email];healthcheck=true;saslQop=auth-conf;auth=KERBEROS;user.principal=spark2x/[hidden email];user.keytab=/opt/huawei/Bigdata/FusionInsight_Spark2x_V100R002C70/install/FusionInsight-Spark2x-2.1/keytab/spark2x/JDBCServer/spark2x.keytab | org.apache.hive.jdbc.HiveConnection.<init>(HiveConnection.java:145) > 2017-06-27 14:06:37,003 | INFO | [main] | Supplied authorities: 10.19.91.225:22550 | org.apache.hive.jdbc.Utils.parseURL(Utils.java:391) > 2017-06-27 14:06:37,625 | WARN | [main] | HiveConf of name hive.thriftProxy.user does not exist | org.apache.hadoop.hive.conf.HiveConf.initialize(HiveConf.java:2920) > 2017-06-27 14:06:37,625 | WARN | [main] | HiveConf of name hive.thriftProxy.address does not exist | org.apache.hadoop.hive.conf.HiveConf.initialize(HiveConf.java:2920) > 2017-06-27 14:06:37,629 | INFO | [main] | Resolved authority: 10.19.91.225:22550 | org.apache.hive.jdbc.Utils.parseURL(Utils.java:490) > 2017-06-27 14:06:37,785 | WARN | [main] | Unable to load native-hadoop library for your platform... using builtin-java classes where applicable | org.apache.hadoop.util.NativeCodeLoader.<clinit>(NativeCodeLoader.java:62) > 2017-06-27 14:06:38,050 | INFO | [main] | user login success. | org.apache.hive.jdbc.HiveConnection.login(HiveConnection.java:295) > 2017-06-27 14:06:38,118 | INFO | [main] | Will try to open client transport with JDBC Uri: jdbc:hive2://10.19.91.225:22550/default;principal=spark2x/[hidden email];healthcheck=true;saslQop=auth-conf;auth=KERBEROS;user.principal=spark2x/[hidden email];user.keytab=/opt/huawei/Bigdata/FusionInsight_Spark2x_V100R002C70/install/FusionInsight-Spark2x-2.1/keytab/spark2x/JDBCServer/spark2x.keytab | org.apache.hive.jdbc.HiveConnection.openTransport(HiveConnection.java:317) > 2017-06-27 14:06:38,177 | INFO | [HiveServer2-Handler-Pool: Thread-258] | Client protocol version: HIVE_CLI_SERVICE_PROTOCOL_V8 | org.apache.hive.service.cli.thrift.ThriftCLIService.OpenSession(ThriftCLIService.java:461) > 2017-06-27 14:06:38,177 | INFO | [HiveServer2-Handler-Pool: Thread-258] | Trying to connect to metastore with URI thrift://10.19.91.224:21088 | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:469) > 2017-06-27 14:06:38,180 | INFO | [HiveServer2-Handler-Pool: Thread-258] | Opened a connection to metastore, current connections: 38 | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:514) > 2017-06-27 14:06:38,181 | INFO | [HiveServer2-Handler-Pool: Thread-258] | Connected to metastore. | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:593) > 2017-06-27 14:06:38,245 | WARN | [HiveServer2-Handler-Pool: Thread-258] | load mapred-default.xml, HIVE_CONF_DIR env not found! | org.apache.hadoop.hive.ql.session.SessionState.loadMapredDefaultXml(SessionState.java:1101) > 2017-06-27 14:06:38,246 | INFO | [HiveServer2-Handler-Pool: Thread-258] | Closed a connection to metastore, current connections: 37 | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.close(HiveMetaStoreClient.java:622) > 2017-06-27 14:06:38,247 | INFO | [HiveServer2-Handler-Pool: Thread-258] | Trying to connect to metastore with URI thrift://10.19.91.224:21088 | org.apache.hadoop.hive.metastore.HiveMetaStoreClient.open(HiveMetaStoreClient.java:469) > Expected Result : Secondary index creation on timestamp column should be success. -- This message was sent by Atlassian JIRA (v6.4.14#64029) |
Free forum by Nabble | Edit this page |