Login  Register

Re: query err 'NullPointerException' but fine after table cached in memory

Posted by Li Peng on Jan 12, 2017; 8:14am
URL: http://apache-carbondata-dev-mailing-list-archive.168.s1.nabble.com/query-err-NullPointerException-but-fine-after-table-cached-in-memory-tp6032p6035.html

Hi,
Executor log as follows:

INFO  12-01 15:45:58,574 - Registered signal handlers for [TERM, HUP, INT]
INFO  12-01 15:45:59,380 - Changing view acls to: yarn,spark
INFO  12-01 15:45:59,387 - Changing modify acls to: yarn,spark
INFO  12-01 15:45:59,387 - SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(yarn, spark); users with modify permissions: Set(yarn, spark)
INFO  12-01 15:46:00,422 - Changing view acls to: yarn,spark
INFO  12-01 15:46:00,422 - Changing modify acls to: yarn,spark
INFO  12-01 15:46:00,422 - SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(yarn, spark); users with modify permissions: Set(yarn, spark)
INFO  12-01 15:46:00,825 - Slf4jLogger started
INFO  12-01 15:46:00,873 - Starting remoting
INFO  12-01 15:46:01,046 - Remoting started; listening on addresses :[akka.tcp://sparkExecutorActorSystem@dpnode03:42230]
INFO  12-01 15:46:01,051 - Successfully started service 'sparkExecutorActorSystem' on port 42230.
INFO  12-01 15:46:01,117 - Created local directory at /data02/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-f1549267-39fd-4f92-a793-db3975c05272
INFO  12-01 15:46:01,117 - Created local directory at /data03/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-e1f964d4-877c-45e2-88a0-3d9d14ebf47e
INFO  12-01 15:46:01,117 - Created local directory at /data04/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-63faa7b7-248b-424a-98bb-ab85d9aef0c3
INFO  12-01 15:46:01,118 - Created local directory at /data05/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-e1ac73f4-54e0-4239-933f-2f16f6997a49
INFO  12-01 15:46:01,118 - Created local directory at /data06/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-5b8bf328-594a-419b-af4d-d34a8a350dce
INFO  12-01 15:46:01,118 - Created local directory at /data07/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-6eed5405-1b71-4d70-931a-3c68a8bff14a
INFO  12-01 15:46:01,119 - Created local directory at /data08/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-5dfcf5a3-d06c-4ae7-a2ec-1a41eb709a26
INFO  12-01 15:46:01,119 - Created local directory at /data09/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-8ed5f4f3-6c92-4f46-adea-362324bed778
INFO  12-01 15:46:01,119 - Created local directory at /data10/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-c7367242-4a68-4b45-9a13-cffd4793b46e
INFO  12-01 15:46:01,119 - Created local directory at /data11/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-2e25390e-835e-4f1e-bca5-96b6ecdc7247
INFO  12-01 15:46:01,120 - Created local directory at /data12/hadoop/yarn/local/usercache/spark/appcache/application_1484187732071_0003/blockmgr-c0911b1e-9c85-4393-bdab-db35a772b4fb
INFO  12-01 15:46:01,141 - MemoryStore started with capacity 5.5 GB
INFO  12-01 15:46:01,632 - Connecting to driver: spark://CoarseGrainedScheduler@192.168.50.8:39206
INFO  12-01 15:46:01,681 - Successfully registered with driver
INFO  12-01 15:46:01,684 - Starting executor ID 2 on host dpnode03
INFO  12-01 15:46:01,722 - Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 47403.
INFO  12-01 15:46:01,723 - Server created on 47403
INFO  12-01 15:46:01,728 - external shuffle service port = 7337
INFO  12-01 15:46:01,728 - Trying to register BlockManager
INFO  12-01 15:46:01,743 - Registered BlockManager
INFO  12-01 15:46:01,747 - Registering executor with local external shuffle service.
INFO  12-01 15:46:01,778 - Using REPL class URI: http://192.168.50.8:44119
INFO  12-01 15:49:19,059 - Got assigned task 0
INFO  12-01 15:49:19,066 - Running task 0.0 in stage 0.0 (TID 0)
INFO  12-01 15:49:19,270 - Started reading broadcast variable 0
INFO  12-01 15:49:19,442 - Block broadcast_0_piece0 stored as bytes in memory (estimated size 18.8 KB, free 18.8 KB)
INFO  12-01 15:49:19,468 - Reading broadcast variable 0 took 198 ms
INFO  12-01 15:49:19,616 - Block broadcast_0 stored as values in memory (estimated size 35.2 KB, free 54.0 KB)
INFO  12-01 15:49:21,375 - *************************/usr/hdp/2.5.0.0-1245/spark/conf/carbon.properties
INFO  12-01 15:49:21,376 - [Executor task launch worker-0][partitionID:sale;queryID:1457355617520266_0] Query will be executed on table: sale
INFO  12-01 15:49:21,379 - [Executor task launch worker-0][partitionID:sale;queryID:1457355617520266_0] Property file path: /usr/hdp/2.5.0.0-1245/spark/conf/carbon.properties
INFO  12-01 15:49:21,379 - [Executor task launch worker-0][partitionID:sale;queryID:1457355617520266_0] ------Using Carbon.properties --------
INFO  12-01 15:49:21,379 - [Executor task launch worker-0][partitionID:sale;queryID:1457355617520266_0] {carbon.number.of.cores.while.loading=6, carbon.number.of.cores.while.compacting=4, carbon.sort.file.buffer.size=20, carbon.inmemory.record.size=120000, carbon.sort.size=500000, carbon.graph.rowset.size=100000, carbon.ddl.base.hdfs.url=/user/spark, carbon.compaction.level.threshold=8,6, carbon.number.of.cores=4, carbon.kettle.home=/usr/hdp/2.5.0.0-1245/spark/carbonlib/carbonplugins, carbon.storelocation=hdfs://julong/carbondata/carbonstore, carbon.enable.auto.load.merge=true, carbon.enableXXHash=true, carbon.sort.intermediate.files.limit=100, carbon.major.compaction.size=1024, carbon.badRecords.location=/opt/Carbon/Spark/badrecords, carbon.use.local.dir=true, carbon.enable.quick.filter=false}
INFO  12-01 15:49:23,999 - pool-19-thread-1 ****************************Total Number Rows In BTREE: 10380500
INFO  12-01 15:49:24,005 - [Executor task launch worker-0][partitionID:sale;queryID:1457355617520266_0] Column cache size not configured. Therefore default behavior will be considered and no LRU based eviction of columns will be done
ERROR 12-01 15:49:24,565 - [Executor task launch worker-0][partitionID:sale;queryID:1457355617520266_0]
java.lang.NullPointerException
        at org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultIterator.intialiseInfos(AbstractDetailQueryResultIterator.java:117)
        at org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultIterator.<init>(AbstractDetailQueryResultIterator.java:107)
        at org.apache.carbondata.scan.result.iterator.DetailQueryResultIterator.<init>(DetailQueryResultIterator.java:43)
        at org.apache.carbondata.scan.executor.impl.DetailQueryExecutor.execute(DetailQueryExecutor.java:39)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<init>(CarbonScanRDD.scala:216)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:192)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)
ERROR 12-01 15:49:24,577 - Exception in task 0.0 in stage 0.0 (TID 0)
java.lang.RuntimeException: Exception occurred in query execution.Please check logs.
        at scala.sys.package$.error(package.scala:27)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<init>(CarbonScanRDD.scala:226)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:192)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)
INFO  12-01 15:49:34,727 - Got assigned task 3
INFO  12-01 15:49:34,727 - Running task 0.3 in stage 0.0 (TID 3)
INFO  12-01 15:49:34,799 - *************************/usr/hdp/2.5.0.0-1245/spark/conf/carbon.properties
INFO  12-01 15:49:34,799 - [Executor task launch worker-0][partitionID:sale;queryID:1457355617520266_0] Query will be executed on table: sale
ERROR 12-01 15:49:34,813 - [Executor task launch worker-0][partitionID:sale;queryID:1457355617520266_0]
java.lang.NullPointerException
        at org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultIterator.intialiseInfos(AbstractDetailQueryResultIterator.java:117)
        at org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultIterator.<init>(AbstractDetailQueryResultIterator.java:107)
        at org.apache.carbondata.scan.result.iterator.DetailQueryResultIterator.<init>(DetailQueryResultIterator.java:43)
        at org.apache.carbondata.scan.executor.impl.DetailQueryExecutor.execute(DetailQueryExecutor.java:39)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<init>(CarbonScanRDD.scala:216)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:192)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)
ERROR 12-01 15:49:34,816 - Exception in task 0.3 in stage 0.0 (TID 3)
java.lang.RuntimeException: Exception occurred in query execution.Please check logs.
        at scala.sys.package$.error(package.scala:27)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<init>(CarbonScanRDD.scala:226)
        at org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(CarbonScanRDD.scala:192)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
        at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
        at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
        at org.apache.spark.scheduler.Task.run(Task.scala:89)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:227)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:745)

Thanks.




kumarvishal09 wrote
Hi,
Can u please share executor log.

-Regards
Kumar Vishal

On Thu, Jan 12, 2017 at 1:59 PM, Li Peng <[hidden email]> wrote:

> Hello,
>
> use carbondata 0.2.0, following is the problem:
>
> Only one column 'store_id' throws NullPointerException  when query, but it
> works fine  when some value or table is cached in memory.
>
> store_id's  type is int, cardinality is 200 Thousand, is configured about
> dictionary and inverted index.
>
> sql:
> select
> order_code,saletype,checkout_date,cashier_code,item_cont,
> invoice_price,giveamt,saleamt
> from store.sale where store_id=299998
>
> error:
> ERROR 12-01 10:40:16,861 - [Executor task launch
> worker-0][partitionID:sale;queryID:1438806645368420_0]
> java.lang.NullPointerException
>         at
> org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultItera
> tor.intialiseInfos(AbstractDetailQueryResultIterator.java:117)
>         at
> org.apache.carbondata.scan.result.iterator.AbstractDetailQueryResultItera
> tor.<init>(AbstractDetailQueryResultIterator.java:107)
>         at
> org.apache.carbondata.scan.result.iterator.DetailQueryResultIterator.<
> init>(DetailQueryResultIterator.java:43)
>         at
> org.apache.carbondata.scan.executor.impl.DetailQueryExecutor.execute(
> DetailQueryExecutor.java:39)
>         at
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<
> init>(CarbonScanRDD.scala:216)
>         at
> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(
> CarbonScanRDD.scala:192)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
> scala:66)
>         at org.apache.spark.scheduler.Task.run(Task.scala:89)
>         at org.apache.spark.executor.Executor$TaskRunner.run(
> Executor.scala:227)
>         at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
>         at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
>         at java.lang.Thread.run(Thread.java:745)
> ERROR 12-01 10:40:16,874 - Exception in task 0.1 in stage 0.0 (TID 1)
> java.lang.RuntimeException: Exception occurred in query execution.Please
> check logs.
>         at scala.sys.package$.error(package.scala:27)
>         at
> org.apache.carbondata.spark.rdd.CarbonScanRDD$$anon$1.<
> init>(CarbonScanRDD.scala:226)
>         at
> org.apache.carbondata.spark.rdd.CarbonScanRDD.compute(
> CarbonScanRDD.scala:192)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)
>         at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:313)
>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:277)
>         at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.
> scala:66)
>         at org.apache.spark.scheduler.Task.run(Task.scala:89)
>         at org.apache.spark.executor.Executor$TaskRunner.run(
> Executor.scala:227)
>         at
> java.util.concurrent.ThreadPoolExecutor.runWorker(
> ThreadPoolExecutor.java:1145)
>         at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(
> ThreadPoolExecutor.java:615)
>         at java.lang.Thread.run(Thread.java:745)
> ------------------------------------------------------------
> -------------------------------------------------------------------
> Same err 'NullPointerException' in following sql:
>
> select * from store.sale where store_id=100000
> select * from store.sale where store_id=100001
> select * from store.sale where store_id=100002
> select * from store.sale where store_id=100006
> select * from store.sale where store_id=100011
> select * from store.sale where store_id=299999
>
> But  fine and can return results in following sql:
>
> select * from store.sale where store_id=100008
> select * from store.sale where store_id=100009
> select * from store.sale where store_id=100010
> select * from store.sale where store_id=100013
> select * from store.sale where store_id=100027
> select * from store.sale limit 10
> select count(*) from store.sale
> select * from store.sale where store_id=100005
> select count(*) from store.sale where store_id=100005
> select distinct(store_id) from store.sale order by store_id
>
>
> But all work fine when table is cached to memory, not throwing
> 'NullPointerException'.
>
>
>
>
> --
> View this message in context: http://apache-carbondata-
> mailing-list-archive.1130556.n5.nabble.com/query-err-
> NullPointerException-but-fine-after-table-cached-in-memory-tp6032.html
> Sent from the Apache CarbonData Mailing List archive mailing list archive
> at Nabble.com.
>