GitHub user kevinjmh opened a pull request:
https://github.com/apache/carbondata/pull/2526 [CARBONDATA-2757][BloomDataMap] Fix bug when building bloomfilter on decimal column 1. support to get raw data from decimal column page when building datamap in loading process 2. convert decimal column to java datatype when rebuilding bloom datamap from query result Be sure to do all of the following checklist to help us incorporate your contribution quickly and easily: - [ ] Any interfaces changed? - [ ] Any backward compatibility impacted? - [ ] Document update required? - [ ] Testing done Please provide details on - Whether new unit test cases have been added or why no new tests are required? - How it is tested? Please attach test report. - Is it a performance related change? Please attach the performance test report. - Any additional information to help reviewers in testing this change. - [ ] For large changes, please consider breaking it into sub-tasks under an umbrella JIRA. You can merge this pull request into a Git repository by running: $ git pull https://github.com/kevinjmh/carbondata bloom_decimal Alternatively you can review and apply these changes as the patch at: https://github.com/apache/carbondata/pull/2526.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #2526 ---- commit 1cdbcf441ec4ae39e5469faad5bc525f8672701d Author: Manhua <kevinjmh@...> Date: 2018-07-19T08:26:18Z fix building bloomfilter datamap on decimal column ---- --- |
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/2526 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7306/ --- |
In reply to this post by qiuchenjian-2
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/2526 Build Failed with Spark 2.2.1, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/6073/ --- |
In reply to this post by qiuchenjian-2
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/2526 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7312/ --- |
In reply to this post by qiuchenjian-2
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/2526 Build Success with Spark 2.2.1, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/6077/ --- |
In reply to this post by qiuchenjian-2
Github user ravipesala commented on the issue:
https://github.com/apache/carbondata/pull/2526 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/5920/ --- |
In reply to this post by qiuchenjian-2
Github user ravipesala commented on the issue:
https://github.com/apache/carbondata/pull/2526 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/5924/ --- |
In reply to this post by qiuchenjian-2
Github user xuchuanyin commented on a diff in the pull request:
https://github.com/apache/carbondata/pull/2526#discussion_r203962823 --- Diff: datamap/bloom/src/main/java/org/apache/carbondata/datamap/bloom/AbstractBloomDataMapWriter.java --- @@ -129,8 +130,12 @@ protected void addValue2BloomIndex(int indexColIdx, Object value) { // convert non-dict dimensions to simple bytes without length // convert internal-dict dimensions to simple bytes without any encode if (indexColumns.get(indexColIdx).isMeasure()) { - if (value == null) { - value = DataConvertUtil.getNullValueForMeasure(indexColumns.get(indexColIdx).getDataType()); + // NULL value of all measures are already processed in `ColumnPage.getData` + // or `RawBytesReadSupport.readRow` with actual data type + + // Carbon stores boolean as byte. Here we convert it for `getValueAsBytes` + if (value instanceof Boolean) { --- End diff -- can we use the datatype as the condition? --- |
In reply to this post by qiuchenjian-2
Github user xuchuanyin commented on a diff in the pull request:
https://github.com/apache/carbondata/pull/2526#discussion_r203962551 --- Diff: core/src/main/java/org/apache/carbondata/core/datastore/page/DecimalColumnPage.java --- @@ -106,4 +109,48 @@ public void setDoublePage(double[] doubleData) { throw new UnsupportedOperationException("invalid data type: " + dataType); } + private BigDecimal getDecimalFromRawData(int rowId) { --- End diff -- Please add comment to describe when to use this method and the below method --- |
In reply to this post by qiuchenjian-2
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/2526 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7351/ --- |
In reply to this post by qiuchenjian-2
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/2526 Build Success with Spark 2.2.1, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/6115/ --- |
In reply to this post by qiuchenjian-2
Github user kevinjmh commented on a diff in the pull request:
https://github.com/apache/carbondata/pull/2526#discussion_r203986298 --- Diff: datamap/bloom/src/main/java/org/apache/carbondata/datamap/bloom/AbstractBloomDataMapWriter.java --- @@ -129,8 +130,12 @@ protected void addValue2BloomIndex(int indexColIdx, Object value) { // convert non-dict dimensions to simple bytes without length // convert internal-dict dimensions to simple bytes without any encode if (indexColumns.get(indexColIdx).isMeasure()) { - if (value == null) { - value = DataConvertUtil.getNullValueForMeasure(indexColumns.get(indexColIdx).getDataType()); + // NULL value of all measures are already processed in `ColumnPage.getData` + // or `RawBytesReadSupport.readRow` with actual data type + + // Carbon stores boolean as byte. Here we convert it for `getValueAsBytes` + if (value instanceof Boolean) { --- End diff -- Yes. FYI, column of boolean type uses column page with inner datatype Byte, and some disagreement on DataType exists between `getData` and `getNull`. I will changed that. --- |
In reply to this post by qiuchenjian-2
Github user kevinjmh commented on a diff in the pull request:
https://github.com/apache/carbondata/pull/2526#discussion_r203986320 --- Diff: core/src/main/java/org/apache/carbondata/core/datastore/page/DecimalColumnPage.java --- @@ -106,4 +109,48 @@ public void setDoublePage(double[] doubleData) { throw new UnsupportedOperationException("invalid data type: " + dataType); } + private BigDecimal getDecimalFromRawData(int rowId) { --- End diff -- Fixed --- |
In reply to this post by qiuchenjian-2
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/2526 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7355/ --- |
In reply to this post by qiuchenjian-2
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/2526 Build Success with Spark 2.2.1, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/6119/ --- |
In reply to this post by qiuchenjian-2
Github user ravipesala commented on the issue:
https://github.com/apache/carbondata/pull/2526 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/5936/ --- |
In reply to this post by qiuchenjian-2
Github user xuchuanyin commented on the issue:
https://github.com/apache/carbondata/pull/2526 LGTM --- |
In reply to this post by qiuchenjian-2
|
Free forum by Nabble | Edit this page |