hi I use the newest version 0.2.0-incubating,and I load 50G data in carbondata.than exec the query below:
2: jdbc:hive2://172.12.1.24:10000> select * from hotel_event_2 where c1 = "key_label_1_10" and c3 > "2005-11-18 00:28:02"; +-----------------+------------------------+------------------------+-------------------------+------------------------+----------------------+----------------------+--------------------------+-----------------------+---------------------+------------------------+--------------------------+-----------------------+------------------------+-------------+------------------------+------------------------+--------+--+ | c1 | c2 | c3 | c4 | c5 | c6 | c7 | c8 | c9 | c10 | c11 | c12 | c13 | c14 | c15 | c16 | c17 | c18 | +-----------------+------------------------+------------------------+-------------------------+------------------------+----------------------+----------------------+--------------------------+-----------------------+---------------------+------------------------+--------------------------+-----------------------+------------------------+-------------+------------------------+------------------------+--------+--+ | key_label_1_10 | 2013-11-27 22:38:27.0 | 2005-11-18 09:28:02.0 | eHOTEL_ssxmc_256067775 | 2011-04-20 21:34:01.0 | eHOTEL_xm_704442173 | eHOTEL_xz_944320369 | eHOTEL_zjlxdm_223506831 | eHOTEL_fjh_491493963 | 513626198103218689 | eHOTEL_lkid_127359383 | eHOTEL_gxdwmc_968918404 | eHOTEL_hylx_12318130 | key_label_2_826173605 | 3296226951 | eHOTEL_ldmc_966801213 | eHOTEL_ldxz_693009883 | other | +-----------------+------------------------+------------------------+-------------------------+------------------------+----------------------+----------------------+--------------------------+-----------------------+---------------------+------------------------+--------------------------+-----------------------+------------------------+-------------+------------------------+------------------------+--------+--+ 1 row selected (5.497 seconds) and than I change the key_label_1_10 to key_label_1_51726972 it throw an error 2: jdbc:hive2://172.12.1.24:10000> select * from hotel_event_2 where c1 = "key_label_1_51726972" and c3 > "2005-11-18 00:28:02"; Error: org.apache.spark.SparkException: Job aborted due to stage failure: Task 3 in stage 0.0 failed 4 times, most recent failure: Lost task 3.3 in stage 0.0 (TID 12, k1222.mlamp.co): java.lang.RuntimeException: Exception occurred in query execution.Please check logs. at scala.sys.package$.error(package.scala:27) at org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<init>(CarbonScanRDD.scala:226) at org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:192) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) at org.apache.spark.scheduler.Task.run(Task.scala:89) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) Driver stacktrace: (state=,code=0) |
Administrator
|
Hi
Are you using hive client to run sql to query carbon table ? -------------------------------------------------------------------- jdbc:hive2://172.12.1.24:10000> select * from hotel_event_2 where c1 = "key_label_1_10" and c3 > "2005-11-18 00:28:02"; Regards Liang
|
In reply to this post by sailingYang
Please provide executor side log
Best Regards
David Cai |
In reply to this post by Liang Chen
yes
1,I load the data to carbondata use spark job,it work successful. 2,I run the thrift server with no exception throw. 3,finally I run the beeline shell to contact thrift server to query. |
This post was updated on .
In reply to this post by David CaiQiang
this is the executor side log,I get this from spark application master.it looks the same...
16/12/26 19:44:44 ERROR rdd.CarbonScanRDD: [Executor task launch worker-1][partitionID:2;queryID:18503694729979314_2] java.lang.NullPointerException at org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultIterator.intialiseInfos(AbstractDetailQueryResultIterator.java:117) at org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultIterator.<init>(AbstractDetailQueryResultIterator.java:107) at org.apache.carbondata.scan.result.iterator.DetailQueryResultIterator.<init>(DetailQueryResultIterator.java:43) at org.apache.carbondata.scan.executor.impl.DetailQueryExecutor.execute(DetailQueryExecutor.java:39) at org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<init>(CarbonScanRDD.scala:216) at org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:192) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38) at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306) at org.apache.spark.rdd.RDD.iterator(RDD.scala:270) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66) at org.apache.spark.scheduler.Task.run(Task.scala:89) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) 16/12/26 19:44:44 ERROR executor.Executor: Exception in task 2.0 in stage 2.0 (TID 15) |
Free forum by Nabble | Edit this page |