Login  Register

same query and I change the value than throw a error

classic Classic list List threaded Threaded
5 messages Options Options
Embed post
Permalink
Reply | Threaded
Open this post in threaded view
| More
Print post
Permalink

same query and I change the value than throw a error

sailingYang
3 posts
hi I use the newest version 0.2.0-incubating,and I load 50G data in carbondata.than exec the query below:
2: jdbc:hive2://172.12.1.24:10000> select * from hotel_event_2 where c1 = "key_label_1_10" and c3 > "2005-11-18 00:28:02";
+-----------------+------------------------+------------------------+-------------------------+------------------------+----------------------+----------------------+--------------------------+-----------------------+---------------------+------------------------+--------------------------+-----------------------+------------------------+-------------+------------------------+------------------------+--------+--+
|   c1   |     c2      |          c3          |      c4       |      c5       |      c6       |      c7       |      c8       |      c9       |         c10         |      c11       |      c12       |      c13      |      c14       |    c15     |      c16       |      c17       |  c18  |
+-----------------+------------------------+------------------------+-------------------------+------------------------+----------------------+----------------------+--------------------------+-----------------------+---------------------+------------------------+--------------------------+-----------------------+------------------------+-------------+------------------------+------------------------+--------+--+
| key_label_1_10  | 2013-11-27 22:38:27.0  | 2005-11-18 09:28:02.0  | eHOTEL_ssxmc_256067775  | 2011-04-20 21:34:01.0  | eHOTEL_xm_704442173  | eHOTEL_xz_944320369  | eHOTEL_zjlxdm_223506831  | eHOTEL_fjh_491493963  | 513626198103218689  | eHOTEL_lkid_127359383  | eHOTEL_gxdwmc_968918404  | eHOTEL_hylx_12318130  | key_label_2_826173605  | 3296226951  | eHOTEL_ldmc_966801213  | eHOTEL_ldxz_693009883  | other  |
+-----------------+------------------------+------------------------+-------------------------+------------------------+----------------------+----------------------+--------------------------+-----------------------+---------------------+------------------------+--------------------------+-----------------------+------------------------+-------------+------------------------+------------------------+--------+--+
1 row selected (5.497 seconds)

and than I change the key_label_1_10 to key_label_1_51726972 it throw an error
2: jdbc:hive2://172.12.1.24:10000> select * from hotel_event_2 where c1 = "key_label_1_51726972" and c3 > "2005-11-18 00:28:02";
Error: org.apache.spark.SparkException: Job aborted due to stage failure: Task 3 in stage 0.0 failed 4 times, most recent failure: Lost task 3.3 in stage 0.0 (TID 12, k1222.mlamp.co): java.lang.RuntimeException: Exception occurred in query execution.Please check logs.
        at scala.sys.package$.error(package.scala:27)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<init>(CarbonScanRDD.scala:226)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:192)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:745)

Driver stacktrace: (state=,code=0)
Reply | Threaded
Open this post in threaded view
| More
Print post
Permalink

Re: same query and I change the value than throw a error

Liang Chen
Administrator
313 posts
Hi

Are you using hive client to run sql to query carbon table ?
--------------------------------------------------------------------
jdbc:hive2://172.12.1.24:10000> select * from hotel_event_2 where c1 = "key_label_1_10" and c3 > "2005-11-18 00:28:02";

Regards
Liang
sailingYang wrote
hi I use the newest version 0.2.0-incubating,and I load 50G data in carbondata.than exec the query below:
2: jdbc:hive2://172.12.1.24:10000> select * from hotel_event_2 where c1 = "key_label_1_10" and c3 > "2005-11-18 00:28:02";
+-----------------+------------------------+------------------------+-------------------------+------------------------+----------------------+----------------------+--------------------------+-----------------------+---------------------+------------------------+--------------------------+-----------------------+------------------------+-------------+------------------------+------------------------+--------+--+
|   c1   |     c2      |          c3          |      c4       |      c5       |      c6       |      c7       |      c8       |      c9       |         c10         |      c11       |      c12       |      c13      |      c14       |    c15     |      c16       |      c17       |  c18  |
+-----------------+------------------------+------------------------+-------------------------+------------------------+----------------------+----------------------+--------------------------+-----------------------+---------------------+------------------------+--------------------------+-----------------------+------------------------+-------------+------------------------+------------------------+--------+--+
| key_label_1_10  | 2013-11-27 22:38:27.0  | 2005-11-18 09:28:02.0  | eHOTEL_ssxmc_256067775  | 2011-04-20 21:34:01.0  | eHOTEL_xm_704442173  | eHOTEL_xz_944320369  | eHOTEL_zjlxdm_223506831  | eHOTEL_fjh_491493963  | 513626198103218689  | eHOTEL_lkid_127359383  | eHOTEL_gxdwmc_968918404  | eHOTEL_hylx_12318130  | key_label_2_826173605  | 3296226951  | eHOTEL_ldmc_966801213  | eHOTEL_ldxz_693009883  | other  |
+-----------------+------------------------+------------------------+-------------------------+------------------------+----------------------+----------------------+--------------------------+-----------------------+---------------------+------------------------+--------------------------+-----------------------+------------------------+-------------+------------------------+------------------------+--------+--+
1 row selected (5.497 seconds)

and than I change the key_label_1_10 to key_label_1_51726972 it throw an error
2: jdbc:hive2://172.12.1.24:10000> select * from hotel_event_2 where c1 = "key_label_1_51726972" and c3 > "2005-11-18 00:28:02";
Error: org.apache.spark.SparkException: Job aborted due to stage failure: Task 3 in stage 0.0 failed 4 times, most recent failure: Lost task 3.3 in stage 0.0 (TID 12, k1222.mlamp.co): java.lang.RuntimeException: Exception occurred in query execution.Please check logs.
        at scala.sys.package$.error(package.scala:27)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<init>(CarbonScanRDD.scala:226)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:192)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:745)

Driver stacktrace: (state=,code=0)
Reply | Threaded
Open this post in threaded view
| More
Print post
Permalink

Re: same query and I change the value than throw a error

David CaiQiang
171 posts
In reply to this post by sailingYang
Please provide executor side log
Best Regards
David Cai
Reply | Threaded
Open this post in threaded view
| More
Print post
Permalink

Re: same query and I change the value than throw a error

sailingYang
3 posts
In reply to this post by Liang Chen
yes
1,I load the data to carbondata use spark job,it work successful.
2,I run the thrift server with no exception throw.
3,finally I run the beeline shell to contact thrift server to query.
Reply | Threaded
Open this post in threaded view
| More
Print post
Permalink

Re: same query and I change the value than throw a error

sailingYang
3 posts
This post was updated on Dec 26, 2016; 11:51am.
In reply to this post by David CaiQiang
this is the executor side log,I get this from spark application master.it looks the same...
16/12/26 19:44:44 ERROR rdd.CarbonScanRDD: [Executor task launch worker-1][partitionID:2;queryID:18503694729979314_2]
java.lang.NullPointerException
        at org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultIterator.intialiseInfos(AbstractDetailQueryResultIterator.java:117)
        at org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultIterator.<init>(AbstractDetailQueryResultIterator.java:107)
        at org.apache.carbondata.scan.result.iterator.DetailQueryResultIterator.<init>(DetailQueryResultIterator.java:43)
        at org.apache.carbondata.scan.executor.impl.DetailQueryExecutor.execute(DetailQueryExecutor.java:39)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<init>(CarbonScanRDD.scala:216)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:192)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:745)
16/12/26 19:44:44 ERROR executor.Executor: Exception in task 2.0 in stage 2.0 (TID 15)