same query and I change the value than throw a error

classic Classic list List threaded Threaded
5 messages Options
Reply | Threaded
Open this post in threaded view
|

same query and I change the value than throw a error

sailingYang
hi I use the newest version 0.2.0-incubating,and I load 50G data in carbondata.than exec the query below:
2: jdbc:hive2://172.12.1.24:10000> select * from hotel_event_2 where c1 = "key_label_1_10" and c3 > "2005-11-18 00:28:02";
+-----------------+------------------------+------------------------+-------------------------+------------------------+----------------------+----------------------+--------------------------+-----------------------+---------------------+------------------------+--------------------------+-----------------------+------------------------+-------------+------------------------+------------------------+--------+--+
|   c1   |     c2      |          c3          |      c4       |      c5       |      c6       |      c7       |      c8       |      c9       |         c10         |      c11       |      c12       |      c13      |      c14       |    c15     |      c16       |      c17       |  c18  |
+-----------------+------------------------+------------------------+-------------------------+------------------------+----------------------+----------------------+--------------------------+-----------------------+---------------------+------------------------+--------------------------+-----------------------+------------------------+-------------+------------------------+------------------------+--------+--+
| key_label_1_10  | 2013-11-27 22:38:27.0  | 2005-11-18 09:28:02.0  | eHOTEL_ssxmc_256067775  | 2011-04-20 21:34:01.0  | eHOTEL_xm_704442173  | eHOTEL_xz_944320369  | eHOTEL_zjlxdm_223506831  | eHOTEL_fjh_491493963  | 513626198103218689  | eHOTEL_lkid_127359383  | eHOTEL_gxdwmc_968918404  | eHOTEL_hylx_12318130  | key_label_2_826173605  | 3296226951  | eHOTEL_ldmc_966801213  | eHOTEL_ldxz_693009883  | other  |
+-----------------+------------------------+------------------------+-------------------------+------------------------+----------------------+----------------------+--------------------------+-----------------------+---------------------+------------------------+--------------------------+-----------------------+------------------------+-------------+------------------------+------------------------+--------+--+
1 row selected (5.497 seconds)

and than I change the key_label_1_10 to key_label_1_51726972 it throw an error
2: jdbc:hive2://172.12.1.24:10000> select * from hotel_event_2 where c1 = "key_label_1_51726972" and c3 > "2005-11-18 00:28:02";
Error: org.apache.spark.SparkException: Job aborted due to stage failure: Task 3 in stage 0.0 failed 4 times, most recent failure: Lost task 3.3 in stage 0.0 (TID 12, k1222.mlamp.co): java.lang.RuntimeException: Exception occurred in query execution.Please check logs.
        at scala.sys.package$.error(package.scala:27)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<init>(CarbonScanRDD.scala:226)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:192)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:745)

Driver stacktrace: (state=,code=0)
Reply | Threaded
Open this post in threaded view
|

Re: same query and I change the value than throw a error

Liang Chen
Administrator
Hi

Are you using hive client to run sql to query carbon table ?
--------------------------------------------------------------------
jdbc:hive2://172.12.1.24:10000> select * from hotel_event_2 where c1 = "key_label_1_10" and c3 > "2005-11-18 00:28:02";

Regards
Liang
sailingYang wrote
hi I use the newest version 0.2.0-incubating,and I load 50G data in carbondata.than exec the query below:
2: jdbc:hive2://172.12.1.24:10000> select * from hotel_event_2 where c1 = "key_label_1_10" and c3 > "2005-11-18 00:28:02";
+-----------------+------------------------+------------------------+-------------------------+------------------------+----------------------+----------------------+--------------------------+-----------------------+---------------------+------------------------+--------------------------+-----------------------+------------------------+-------------+------------------------+------------------------+--------+--+
|   c1   |     c2      |          c3          |      c4       |      c5       |      c6       |      c7       |      c8       |      c9       |         c10         |      c11       |      c12       |      c13      |      c14       |    c15     |      c16       |      c17       |  c18  |
+-----------------+------------------------+------------------------+-------------------------+------------------------+----------------------+----------------------+--------------------------+-----------------------+---------------------+------------------------+--------------------------+-----------------------+------------------------+-------------+------------------------+------------------------+--------+--+
| key_label_1_10  | 2013-11-27 22:38:27.0  | 2005-11-18 09:28:02.0  | eHOTEL_ssxmc_256067775  | 2011-04-20 21:34:01.0  | eHOTEL_xm_704442173  | eHOTEL_xz_944320369  | eHOTEL_zjlxdm_223506831  | eHOTEL_fjh_491493963  | 513626198103218689  | eHOTEL_lkid_127359383  | eHOTEL_gxdwmc_968918404  | eHOTEL_hylx_12318130  | key_label_2_826173605  | 3296226951  | eHOTEL_ldmc_966801213  | eHOTEL_ldxz_693009883  | other  |
+-----------------+------------------------+------------------------+-------------------------+------------------------+----------------------+----------------------+--------------------------+-----------------------+---------------------+------------------------+--------------------------+-----------------------+------------------------+-------------+------------------------+------------------------+--------+--+
1 row selected (5.497 seconds)

and than I change the key_label_1_10 to key_label_1_51726972 it throw an error
2: jdbc:hive2://172.12.1.24:10000> select * from hotel_event_2 where c1 = "key_label_1_51726972" and c3 > "2005-11-18 00:28:02";
Error: org.apache.spark.SparkException: Job aborted due to stage failure: Task 3 in stage 0.0 failed 4 times, most recent failure: Lost task 3.3 in stage 0.0 (TID 12, k1222.mlamp.co): java.lang.RuntimeException: Exception occurred in query execution.Please check logs.
        at scala.sys.package$.error(package.scala:27)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<init>(CarbonScanRDD.scala:226)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:192)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:745)

Driver stacktrace: (state=,code=0)
Reply | Threaded
Open this post in threaded view
|

Re: same query and I change the value than throw a error

David CaiQiang
In reply to this post by sailingYang
Please provide executor side log
Best Regards
David Cai
Reply | Threaded
Open this post in threaded view
|

Re: same query and I change the value than throw a error

sailingYang
In reply to this post by Liang Chen
yes
1,I load the data to carbondata use spark job,it work successful.
2,I run the thrift server with no exception throw.
3,finally I run the beeline shell to contact thrift server to query.
Reply | Threaded
Open this post in threaded view
|

Re: same query and I change the value than throw a error

sailingYang
This post was updated on .
In reply to this post by David CaiQiang
this is the executor side log,I get this from spark application master.it looks the same...
16/12/26 19:44:44 ERROR rdd.CarbonScanRDD: [Executor task launch worker-1][partitionID:2;queryID:18503694729979314_2]
java.lang.NullPointerException
        at org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultIterator.intialiseInfos(AbstractDetailQueryResultIterator.java:117)
        at org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultIterator.<init>(AbstractDetailQueryResultIterator.java:107)
        at org.apache.carbondata.scan.result.iterator.DetailQueryResultIterator.<init>(DetailQueryResultIterator.java:43)
        at org.apache.carbondata.scan.executor.impl.DetailQueryExecutor.execute(DetailQueryExecutor.java:39)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<init>(CarbonScanRDD.scala:216)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:192)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:306)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:270)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:745)
16/12/26 19:44:44 ERROR executor.Executor: Exception in task 2.0 in stage 2.0 (TID 15)