xubo245 created CARBONDATA-2661:
----------------------------------- Summary: Query failed with group by column when MV Datamap created without group by column Key: CARBONDATA-2661 URL: https://issues.apache.org/jira/browse/CARBONDATA-2661 Project: CarbonData Issue Type: Bug Reporter: xubo245 Assignee: xubo245 {code:java} drop table if exists fact_table1; CREATE TABLE fact_table1 (empno int, empname String, designation String, doj Timestamp, workgroupcategory int, workgroupcategoryname String, deptno int, deptname String, projectcode int, projectjoindate Timestamp, projectenddate Timestamp,attendance int, utilization int,salary int) STORED BY 'org.apache.carbondata.format'; LOAD DATA local inpath 'hdfs://hacluster/user/hive/warehouse//data_mv.csv' INTO TABLE fact_table1 OPTIONS('DELIMITER'= ',', 'QUOTECHAR'= '"','timestampformat'='dd-MM-yyyy'); LOAD DATA local inpath 'hdfs://hacluster/user/hive/warehouse//data_mv.csv' INTO TABLE fact_table1 OPTIONS('DELIMITER'= ',', 'QUOTECHAR'= '"','timestampformat'='dd-MM-yyyy'); 0: jdbc:hive2://hadoop1:10000> create datamap mv1 using 'mv' as select sum(salary),count(empno) from fact_table1 group by empname; +---------+--+ | Result | +---------+--+ +---------+--+ No rows selected (0.399 seconds) 0: jdbc:hive2://hadoop1:10000> rebuild datamap mv1; +---------+--+ | Result | +---------+--+ +---------+--+ No rows selected (1.57 seconds) 0: jdbc:hive2://hadoop1:10000> create datamap mv2 using 'mv' as select sum(salary) from fact_table1 group by empname; +---------+--+ | Result | +---------+--+ +---------+--+ No rows selected (0.321 seconds) 0: jdbc:hive2://hadoop1:10000> rebuild datamap mv2; +---------+--+ | Result | +---------+--+ +---------+--+ No rows selected (1.241 seconds) 0: jdbc:hive2://hadoop1:10000> show datamap on table fact_table1; +--------------+------------+--------------------------+--+ | DataMapName | ClassName | Associated Table | +--------------+------------+--------------------------+--+ | datamap25 | mv | default.datamap25_table | | mv1 | mv | default.mv1_table | | mv2 | mv | default.mv2_table | +--------------+------------+--------------------------+--+ 3 rows selected (0.047 seconds) 0: jdbc:hive2://hadoop1:10000> select * from default.mv1_table; +-------------+--------------+--+ | sum_salary | count_empno | +-------------+--------------+--+ | 172332 | 18 | | 162972 | 18 | | 90720 | 18 | | 202572 | 18 | | 90720 | 18 | | 128232 | 18 | | 130410 | 18 | | 202464 | 18 | | 243846 | 18 | | 238410 | 18 | +-------------+--------------+--+ 10 rows selected (0.314 seconds) 0: jdbc:hive2://hadoop1:10000> select count(*) from default.mv1_table; +-----------+--+ | count(1) | +-----------+--+ | 10 | +-----------+--+ 1 row selected (0.139 seconds) 0: jdbc:hive2://hadoop1:10000> select * from default.mv2_table; +-------------+--+ | sum_salary | +-------------+--+ | 172332 | | 162972 | | 90720 | | 202464 | | 243846 | | 128232 | | 130410 | | 90720 | | 202572 | | 238410 | +-------------+--+ 10 rows selected (0.262 seconds) 0: jdbc:hive2://hadoop1:10000> explain select sum(salary) from fact_table1 group by empname; +------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+--+ | plan | +------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+--+ | == CarbonData Profiler == Table Scan on fact_table1 - total blocklets: 2 - filter: none - pruned by Main DataMap - skipped blocklets: 0 | | == Physical Plan == *HashAggregate(keys=[empname#2499], functions=[sum(cast(salary#2511 as bigint))]) +- Exchange hashpartitioning(empname#2499, 200) +- *HashAggregate(keys=[empname#2499], functions=[partial_sum(cast(salary#2511 as bigint))]) +- *BatchedScan CarbonDatasourceHadoopRelation [ Database name :default, Table name :fact_table1, Schema :Some(StructType(StructField(empno,IntegerType,true), StructField(empname,StringType,true), StructField(designation,StringType,true), StructField(doj,TimestampType,true), StructField(workgroupcategory,IntegerType,true), StructField(workgroupcategoryname,StringType,true), StructField(deptno,IntegerType,true), StructField(deptname,StringType,true), StructField(projectcode,IntegerType,true), StructField(projectjoindate,TimestampType,true), StructField(projectenddate,TimestampType,true), StructField(attendance,IntegerType,true), StructField(utilization,IntegerType,true), StructField(salary,IntegerType,true))) ] default.fact_table1[empname#2499,salary#2511] | +------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+--+ 2 rows selected (0.164 seconds) 0: jdbc:hive2://hadoop1:10000> explain select empname,sum(salary) from fact_table1 group by empname; Error: org.apache.spark.sql.AnalysisException: cannot resolve '`fact_table1.empname`' given input columns: [sum_salary, count_empno]; line 6 pos 11; 'Aggregate ['fact_table1.empname], ['fact_table1.empname, 'sum('gen_subsumer_0.sum(salary)) AS sum(salary)#2802] +- 'SubqueryAlias gen_subsumer_0 +- 'Aggregate ['fact_table1.empname], [sum(sum_salary#2495L) AS sum(salary)#2800L, sum(count_empno#2496L) AS count(empno)#2801L] +- SubqueryAlias mv1_table +- Relation[sum_salary#2495L,count_empno#2496L] CarbonDatasourceHadoopRelation [ Database name :default, Table name :mv1_table, Schema :Some(StructType(StructField(sum_salary,LongType,true), StructField(count_empno,LongType,true))) ] (state=,code=0) 0: jdbc:hive2://hadoop1:10000> create datamap mv3 using 'mv' as select empname,sum(salary) from fact_table1 group by empname; +---------+--+ | Result | +---------+--+ +---------+--+ No rows selected (0.318 seconds) 0: jdbc:hive2://hadoop1:10000> rebuild datamap mv3; +---------+--+ | Result | +---------+--+ +---------+--+ No rows selected (1.437 seconds) 0: jdbc:hive2://hadoop1:10000> drop datamap mv3; +---------+--+ | Result | +---------+--+ +---------+--+ No rows selected (0.941 seconds) 0: jdbc:hive2://hadoop1:10000> create datamap mv3 using 'mv' as select empname,sum(salary),count(empno) from fact_table1 group by empname; +---------+--+ | Result | +---------+--+ +---------+--+ No rows selected (0.537 seconds) 0: jdbc:hive2://hadoop1:10000> create datamap mv4 using 'mv' as select empname,sum(salary) from fact_table1 group by empname; +---------+--+ | Result | +---------+--+ +---------+--+ No rows selected (0.271 seconds) 0: jdbc:hive2://hadoop1:10000> rebuild datamap mv3; +---------+--+ | Result | +---------+--+ +---------+--+ No rows selected (1.525 seconds) 0: jdbc:hive2://hadoop1:10000> rebuild datamap mv4; +---------+--+ | Result | +---------+--+ +---------+--+ No rows selected (0.993 seconds) 0: jdbc:hive2://hadoop1:10000> explain select empname,sum(salary) from fact_table1 group by empname; Error: org.apache.spark.sql.AnalysisException: cannot resolve '`fact_table1.empname`' given input columns: [sum_salary, count_empno]; line 6 pos 11; 'Aggregate ['fact_table1.empname], ['fact_table1.empname, 'sum('gen_subsumer_0.sum(salary)) AS sum(salary)#3113] +- 'SubqueryAlias gen_subsumer_0 +- 'Aggregate ['fact_table1.empname], [sum(sum_salary#2495L) AS sum(salary)#3111L, sum(count_empno#2496L) AS count(empno)#3112L] +- SubqueryAlias mv1_table +- Relation[sum_salary#2495L,count_empno#2496L] CarbonDatasourceHadoopRelation [ Database name :default, Table name :mv1_table, Schema :Some(StructType(StructField(sum_salary,LongType,true), StructField(count_empno,LongType,true))) ] (state=,code=0) {code} -- This message was sent by Atlassian JIRA (v7.6.3#76005) |
Free forum by Nabble | Edit this page |