manoj mathpal created CARBONDATA-1012:
----------------------------------------- Summary: Decimal value is not supported when we select the Query with integer data type. Key: CARBONDATA-1012 URL: https://issues.apache.org/jira/browse/CARBONDATA-1012 Project: CarbonData Issue Type: Bug Components: sql Affects Versions: 1.1.0 Environment: SPARK 2.1 Reporter: manoj mathpal Attachments: Test_Data1.csv, Test_Data1_h1.csv Steps to reproduces: 1. Create table: create table Test_Boundary_h3 (c1_int int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string string,c6_Timestamp Timestamp,c7_Datatype_Desc string) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' ; 2.Load data into table: load data local inpath '/home/manoj/Downloads/TestData/Data/Test_Data1_h1.csv' OVERWRITE INTO TABLE Test_Boundary_h3 ; Execute Query: select c1_int from Test_Boundary_h3 where c1_int in (2.147483647E9,2345.0,1234.0) ; Result: +-------------+--+ | c1_int | +-------------+--+ | 1234 | | 2345 | | 2147483647 | | 2147483647 | +-------------+--+ ............................................................................................................................. Create table in Carbondata.. 1.Create table: create table Test_Boundary (c1_int int,c2_Bigint Bigint,c3_Decimal Decimal(38,30),c4_double double,c5_string string,c6_Timestamp Timestamp,c7_Datatype_Desc string) STORED BY 'org.apache.carbondata.format' ; 2.Load data into table: LOAD DATA INPATH 'HDFS_URL/BabuStore/Data/Test_Data1.csv' INTO table Test_Boundary OPTIONS('DELIMITER'=',','QUOTECHAR'='','BAD_RECORDS_ACTION'='FORCE','FILEHEADER'='') ; 3:Execute Query: select c1_int from Test_Boundary where c1_int in (2.147483647E9,2345.0,1234.0) ; Result: Error: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 96.0 failed 1 times, most recent failure: Lost task 0.0 in stage 96.0 (TID 302, localhost, executor driver): org.apache.spark.util.TaskCompletionListenerException: java.util.concurrent.ExecutionException: org.apache.carbondata.core.scan.expression.exception.FilterUnsupportedException at org.apache.spark.TaskContextImpl.markTaskCompleted(TaskContextImpl.scala:105) at org.apache.spark.scheduler.Task.run(Task.scala:112) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:282) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) Driver stacktrace: (state=,code=0) Mismatched behavior in HIve and carbondata. -- This message was sent by Atlassian JIRA (v6.3.15#6346) |
Free forum by Nabble | Edit this page |