GitHub user sgururajshetty opened a pull request:
https://github.com/apache/carbondata/pull/2603 [Documentation] Editorial review comment fixed Minor issues fixed (spelling, syntax, and missing info) You can merge this pull request into a Git repository by running: $ git pull https://github.com/sgururajshetty/carbondata editorial_review1 Alternatively you can review and apply these changes as the patch at: https://github.com/apache/carbondata/pull/2603.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #2603 ---- commit 529f80dda6db3ce34e0baf766b03a9a13190b286 Author: sgururajshetty <sgururajshetty@...> Date: 2018-07-25T12:44:07Z Documentation for support for COLUMN_META_CACHE in create table and alter table properties commit d816aaa7a89155b3579906f960ed6a0ba4d4a59f Author: sgururajshetty <sgururajshetty@...> Date: 2018-07-25T12:48:43Z Documentation to support for CACHE_LEVEL in create table and alter table properties commit 8ac243f8e9cff8359b6064352deb823eda7b9835 Author: sgururajshetty <sgururajshetty@...> Date: 2018-07-25T13:24:52Z Review comment fixed commit 98501d35cfd110bcb9e75eb02628f3bce0c0f4ab Author: sgururajshetty <sgururajshetty@...> Date: 2018-07-25T13:26:58Z review comment fixed commit 62caf822cbcde1e519501c1d5db3c5cfc05fbd63 Author: Indhumathi27 <indhumathim27@...> Date: 2018-07-21T10:46:21Z [CARBONDATA-2606]Fix Complex array Pushdown and block auto merge compaction 1.Check for if Complex Column contains ArrayType at n levels and add parent to projection if contains array. 2.Block Auto merge compaction for table containing complex datatype columns. 3.Fix Decimal Datatype scale and precision with two level struct type 4.Fix Dictionary Include for ComplexDataType - If other complex columns other than first complex column is given in dictionary include, then its insertion fails. 5.Fix BadRecord and dateformat for Complex primitive type-DATE This closes #2535 commit d287a102b5c96e54261ac00c77038a1a56161fe9 Author: kumarvishal09 <kumarvishal1802@...> Date: 2018-07-24T14:40:54Z [CARBONDATA-2779]Fixed filter query issue in case of V1/v2 format store Problem: Filter query is failing for V1/V2 carbondata store Root Cause: in V1 store measure min max was not added in blockminmaxindex in executor when filter is applied min max pruning is failing with array index out of cound exception Solution: Need to add min max for measure column same as already handled in driver block pruning This closes #2550 commit b08745f68624ff066e0b23a41ce12d4a99618ac5 Author: Manhua <kevinjmh@...> Date: 2018-07-25T08:51:49Z [CARBONDATA-2783][BloomDataMap][Doc] Update document for bloom filter datamap add example for enable/disable datamap This closes #2554 commit 964d26866468df6be130e9d65d339439cb4cf3b0 Author: praveenmeenakshi56 <praveenmeenakshi56@...> Date: 2018-07-25T15:31:37Z [CARBONDATA-2750] Added Documentation for Local Dictionary Support Added Documentation for Local Dictionary Support This closes #2520 commit 1fa9f64d70123d0bc988427a34c0750283f5daae Author: BJangir <babulaljangir111@...> Date: 2018-07-23T16:44:12Z [CARBONDATA-2772] Size based dictionary fallback is failing even threshold is not reached. Issue:- Size Based Fallback happened even threshold is not reached. RootCause:- Current size calculation is wrong. it is calculated for each data. instead of generated dictionary data . Solution :- Current size should be calculated only for generated dictionary data. This closes #2542 commit eae5817e56a20aecb7694c8d387dbb05b96e1045 Author: kunal642 <kunalkapoor642@...> Date: 2018-07-24T10:42:54Z [CARBONDATA-2778]Fixed bug when select after delete and cleanup is showing empty records Problem: In case if delete operation when it is found that the data being deleted is leading to a state where one complete block data is getting deleted. In that case the status if that block is marked for delete and during the next delete operation run the block is deleted along with its carbonIndex file. The problem arises due to deletion of carbonIndex file because for multiple blocks there can be one carbonIndex file as one carbonIndex file represents one task. Solution: Do not delete the carbondata and carbonIndex file. After compaction it will automatically take care of deleting the stale data and stale segments. This closes #2548 commit 6d6874a11482a8aa79f2280f6572e84b5e3cbc93 Author: dhatchayani <dhatcha.official@...> Date: 2018-07-25T09:11:58Z [CARBONDATA-2753][Compatibility] Row count of page is calculated wrong for old store(V2 store) Row count of page is calculated wrong for V2 store. commit b6f5af6af96140876ec10ff09c3313d9b35ceb36 Author: Sssan520 <liangap2008@...> Date: 2018-07-25T11:36:00Z [CARBONDATA-2782]delete dead code in class 'CarbonCleanFilesCommand' The variables(dmsãindexDms) in function processMetadata are nerver used. This closes #2557 commit cf6bb46d89a9630209ef7f7e797ca8e156369206 Author: ajantha-bhat <ajanthabhat@...> Date: 2018-07-24T07:03:47Z [CARBONDATA-2775] Adaptive encoding fails for Unsafe OnHeap. if, target datatype is SHORT_INT problem: [CARBONDATA-2775] Adaptive encoding fails for Unsafe OnHeap if, target data type is SHORT_INT solution: If ENABLE_OFFHEAP_SORT = false, in carbon property. UnsafeFixLengthColumnPage.java will use different compress logic. Not the raw compression. In that case, for SHORT_INT data type , conversion need to handle. This closes #2546 commit 1fd3889cfec902e30cdcb9af6ed7b7f0df0f20e4 Author: mohammadshahidkhan <mohdshahidkhan1987@...> Date: 2018-07-25T06:44:38Z [HOTFIX] Fixed random test failure Fixed random test failure This closes #2553 commit 0f1aabb87b81670554188b698fc9fb4c598d8942 Author: Jatin <jatin.demla@...> Date: 2018-07-25T19:12:50Z [HOTFIX] CreateDataMapPost Event was skipped in case of preaggregate datamap CreateDataMapPost Event was skipped in case of preaggregate datamap This closes #2562 commit 9b05edbace9ccca5c7def6d5e4cc157ed0eb0d66 Author: ajantha-bhat <ajanthabhat@...> Date: 2018-07-25T19:05:36Z [CARBONDATA-2784][CARBONDATA-2786][SDK writer] Fixed:Forever blocking wait with more than 21 batch of data problem: [CARBONDATA-2784] [SDK writer] Forever blocking wait with more than 21 batch of data, when consumer is dead due to data loading exception (bad record / out of memory) root cause: When the consumer is dead due to data loading exception, writer will be forcefully closed. but queue.clear() cleared only snapshot of entries (10 batches) and close is set to true after that. In between clear() and close = true, If more than 10 batches of data is again put into queue. For 11th batch, queue.put() goes for forever block as consumer is dead. Solution: set close = true, before clearing the queue. This will avoid adding more batches to queue from write(). problem [CARBONDATA-2786] NPE when SDK writer tries to write a file solution and cause: #2387 , in CarbonProperties.java After systemLocation = getStorePath(); Null validation missing for systemLocation. because this can be null in SDK case. As Store location is not applicable for SDK. All a null validation. This closes #2561 commit 1c620f83e3f5b1adb28b73f38c40021b2f06e1d0 Author: Indhumathi27 <indhumathim27@...> Date: 2018-07-27T06:52:25Z [CARBONDATA-2791]Fix Encoding for Double if exceeds LONG.Max_value If Factor(decimalcount) * absMaxValue exceeds LONG.MAX_VALUE, then go for direct compression. This closes #2569 commit 39773ccf01345111bc950d8411646b886111f897 Author: manishgupta88 <tomanishgupta18@...> Date: 2018-07-27T08:41:11Z [HOTFIX] Removed file existence check to improve dataMap loading performance Problem DataMap loading performance degraded after adding file existence check. Analysis When carbonIndex file is read and carbondata file path to its metadata Info map is prepared, file physical existence is getting checked every time which in case of HDFS file system is a namenode call. This degrades the dataMap loading performance. This is done to avoid failures for Handle IUD scenario where after delete operation carbondata file is deleted but the entry still exists in index file. Fix Modified code to check for physical file existence only in case when any IUD operation has happened on the table This closes #2560 commit a55722c1f6dad54d58e7cd69b9f48bfb3693c2a0 Author: Indhumathi27 <indhumathim27@...> Date: 2018-07-27T18:09:50Z [CARBONDATA-2794]Distinct count fails on ArrayOfStruct This PR fixes Code Generator Error thrown when Select filter contains more than one count of distinct of ArrayofStruct with group by Clause This closes #2573 commit 6c4a2056f25ddc61d0fa0b205cb65a95c4298549 Author: mohammadshahidkhan <mohdshahidkhan1987@...> Date: 2018-07-17T11:29:35Z [CARBONDATA-2749][dataload] In HDFS Empty tablestatus file is written during datalaod, iud or compaction when disk is full. Problem: When a failure happens due to disk full during load, IUD or Compaction, then while updating the tablestatus file, the tablestaus.tmp file during atomic file operation remains empty, and in the finally block the empty tablestaus.tmp file is getting renamed to the actual file. This leads to empty tablestatus file. Once such problem happens the tablestatus file can not be retrieved and the already loaded data can be used. Solution: If the failures happens during write then the the schema rename in the finally block must be avoided. This closes #2517 commit 0a5431c1b645af8820486a55dd33530583ea9ae1 Author: Jimmy Casey <casey.jimmy@...> Date: 2018-07-29T21:35:27Z Fixed Spelling Fixed Spelling This closes #2584 commit 6fb852ad86cc7ee7dd97af211a6382ec8f4eb750 Author: chenliang613 <chenliang613@...> Date: 2018-07-27T02:51:45Z [CARBONDATA-2789] Support Hadoop 2.8.3 eco-system integration Add hadoop 2.8.3 profile and passed the compile This closes #2566 commit 937f68cf91bf6708de5c51fe787ea7672b3ee126 Author: ravipesala <ravi.pesala@...> Date: 2018-07-30T12:45:32Z [CARBONDATA-2801]Added documentation for flat folder [CARBONDATA-2801]Added documentation for flat folder This closes #2582 commit e1abff71c7feba90be0ab338ec3b5aabf9cf9380 Author: ajantha-bhat <ajanthabhat@...> Date: 2018-07-25T13:51:02Z [CARBONDATA-2606][Complex DataType Enhancements]Fix Null result if projection column have null primitive column and struct Problem: In case if the actual value of the primitive data type is null, by PR#2489, we are moving all the null values to the end of the collected row without considering the data type. Solution: Place null in the end of output iff the null value is of complex primitive column. This closes #2559 commit 917b230b57cff9c350524c81d939b06af6a35cde Author: akashrn5 <akashnilugal@...> Date: 2018-07-27T07:01:48Z [CARBONDATA-2585]disable local dictionary by default make local dictionary false by default This closes #2570 commit 5d2d6cf053accdd76bf26943db0a0f09dfc492e1 Author: xuchuanyin <xuchuanyin@...> Date: 2018-07-26T15:22:58Z [HotFix][CARBONDATA-2788][BloomDataMap] Fix bugs in incorrect query result with bloom datamap This PR solve two problems which will affect the correctness of the query on bloom. Revert PR2539 After review the code, we found that modification in PR2539 is not needed, so we revert that PR. Bugs in overflow for blocklet count Carbondata stores blocklet count for each block in byte data type, when a block contains more than 128 blocklets, it will overflow the byte limits. Here we change the data type to short. For cache_leve=block, after pruning by main BlockDataMap, the blockletNo in Blocklet is -1, which indicate that the following percedure will scan the whole block -- all the blocklets in the block. So, when doing intersection with the pruned result from BloomDataMap, we need to take care of these blocklets. In this implementation, we added the result from BloomDataMap based on the blocklet's existence in BlockDataMap. This closes #2565 commit ec73f91817062495b085a231279f5133e44b1af8 Author: Indhumathi27 <indhumathim27@...> Date: 2018-07-30T08:48:44Z [CARBONDATA-2798] Fix Dictionary_Include for ComplexDataType Problem1: Select Filter is throwing BufferUnderFlow Exception as cardinality is filled for Non-Dictionary columns. Solution: Check if a complex column has Encoding => Dictionary and fill cardinality for that column only. Problem2: Transactional Table is throwing NullPointerException if csv fileheader is not proper. Solution: Throw CarbonDataLoadingException if csv fileheader is not proper. This closes #2578 commit 6fc498055801ab6cf5181b24ee4db47c0510f172 Author: akashrn5 <akashnilugal@...> Date: 2018-07-30T13:52:29Z [CARBONDATA-2805] Fix the ordering mismatch of segment numbers during cutom compaction Problem: when we have segments from 0 to 6 and i give 1, 2, 3 for custom compaction, then it should create 1.1 as compacted segment, but sometimes it will create 3.1 as compacted segment which is wrong. This is beacuse custom Segment IDs were passing in hashset and finally inserted in hashmap, while identifying segments to be merged. hashmap and hashset does not guarantee the insertion order which may lead to missmatch of segment numbers. Solution: Use LinkedHashSet and LinkedHashMap which always sure about the insertion order. This closes #2585 commit 2f828ce2f438de67dabf565be25b9bbad76b49fc Author: rahul <rahul.kumar@...> Date: 2018-07-04T14:01:51Z [CARBONDATA-2625] While BlockletDataMap loading, avoid multiple times listing of files CarbonReader is very slow for many files as blockletDataMap lists files of folder while loading each segment. This optimization lists once across segment loads. This closes #2441 commit 436c8642185c7d262582c55c6c0a077922125be5 Author: xuchuanyin <xuchuanyin@...> Date: 2018-07-27T03:54:21Z [CARBONDATA-2790][BloomDataMap]Optimize default parameter for bloomfilter datamap To provide better query performance for bloomfilter datamap by default, we optimize bloom_size from 32000 to 640000 and optimize bloom_fpp from 0.01 to 0.00001. This closes #2567 ---- --- |
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/2603 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7739/ --- |
In reply to this post by qiuchenjian-2
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/2603 Build Failed with Spark 2.2.1, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/6465/ --- |
In reply to this post by qiuchenjian-2
Github user brijoobopanna commented on the issue:
https://github.com/apache/carbondata/pull/2603 retest this please --- |
In reply to this post by qiuchenjian-2
Github user ravipesala commented on the issue:
https://github.com/apache/carbondata/pull/2603 SDV Build Fail , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/6129/ --- |
In reply to this post by qiuchenjian-2
Github user chetandb commented on a diff in the pull request:
https://github.com/apache/carbondata/pull/2603#discussion_r207457391 --- Diff: docs/datamap/bloomfilter-datamap-guide.md --- @@ -44,7 +44,7 @@ A Bloom filter is a space-efficient probabilistic data structure that is used to Carbondata introduce BloomFilter as an index datamap to enhance the performance of querying with precise value. --- End diff -- Change introduce to introduced --- |
In reply to this post by qiuchenjian-2
Github user chetandb commented on a diff in the pull request:
https://github.com/apache/carbondata/pull/2603#discussion_r207459790 --- Diff: docs/sdk-guide.md --- @@ -351,7 +351,7 @@ public CarbonWriter buildWriterForCSVInput() throws IOException, InvalidLoadOpti * @throws IOException * @throws InvalidLoadOptionException */ -public CarbonWriter buildWriterForAvroInput() throws IOException, InvalidLoadOptionException; +public CarbonWriter buildWriterForAvroInput(org.apache.avro.Schema schema) throws IOException, InvalidLoadOptionException; ``` --- End diff -- TestSdkJson example code needs to be corrected. testJsonSdkWriter should be static and IOException should be handled import java.io.IOException; import org.apache.carbondata.common.exceptions.sql.InvalidLoadOptionException; import org.apache.carbondata.core.metadata.datatype.DataTypes; import org.apache.carbondata.core.util.CarbonProperties; import org.apache.carbondata.sdk.file.CarbonWriter; import org.apache.carbondata.sdk.file.CarbonWriterBuilder; import org.apache.carbondata.sdk.file.Field; import org.apache.carbondata.sdk.file.Schema; public class TestSdkJson { public static void main(String[] args) throws InvalidLoadOptionException { testJsonSdkWriter(); } public void testJsonSdkWriter() throws InvalidLoadOptionException { String path = "./target/testJsonSdkWriter"; Field[] fields = new Field[2]; fields[0] = new Field("name", DataTypes.STRING); fields[1] = new Field("age", DataTypes.INT); Schema CarbonSchema = new Schema(fields); CarbonWriterBuilder builder = CarbonWriter.builder().outputPath(path); // initialize json writer with carbon schema CarbonWriter writer = builder.buildWriterForJsonInput(CarbonSchema); // one row of json Data as String String JsonRow = "{\"name\":\"abcd\", \"age\":10}"; int rows = 5; for (int i = 0; i < rows; i++) { writer.write(JsonRow); } writer.close(); } } 8.2 --- |
In reply to this post by qiuchenjian-2
Github user chetandb commented on a diff in the pull request:
https://github.com/apache/carbondata/pull/2603#discussion_r207460915 --- Diff: docs/configuration-parameters.md --- @@ -140,7 +140,7 @@ This section provides the details of all the configurations required for CarbonD | carbon.enableMinMax | true | Min max is feature added to enhance query performance. To disable this feature, set it false. | | carbon.dynamicallocation.schedulertimeout | 5 | Specifies the maximum time (unit in seconds) the scheduler can wait for executor to be active. Minimum value is 5 sec and maximum value is 15 sec. | | carbon.scheduler.minregisteredresourcesratio | 0.8 | Specifies the minimum resource (executor) ratio needed for starting the block distribution. The default value is 0.8, which indicates 80% of the requested resource is allocated for starting block distribution. The minimum value is 0.1 min and the maximum value is 1.0. | -| carbon.search.enabled | false | If set to true, it will use CarbonReader to do distributed scan directly instead of using compute framework like spark, thus avoiding limitation of compute framework like SQL optimizer and task scheduling overhead. | +| carbon.search.enabled (Alpha Feature) | false | If set to true, it will use CarbonReader to do distributed scan directly instead of using compute framework like spark, thus avoiding limitation of compute framework like SQL optimizer and task scheduling overhead. | * **Global Dictionary Configurations** --- End diff -- In S3 section. 1. there should not be any Space in parameter . Should be carbon.storelocation. 2. "Concurrent queries are not supported" should be changed to "Only concurrent put (data management operations like load,insert,update)are supported." 3. The "Another way of setting the authentication parameters is as follows" should be removed. --- |
In reply to this post by qiuchenjian-2
Github user chetandb commented on a diff in the pull request:
https://github.com/apache/carbondata/pull/2603#discussion_r207462001 --- Diff: docs/configuration-parameters.md --- @@ -140,7 +140,7 @@ This section provides the details of all the configurations required for CarbonD | carbon.enableMinMax | true | Min max is feature added to enhance query performance. To disable this feature, set it false. | | carbon.dynamicallocation.schedulertimeout | 5 | Specifies the maximum time (unit in seconds) the scheduler can wait for executor to be active. Minimum value is 5 sec and maximum value is 15 sec. | | carbon.scheduler.minregisteredresourcesratio | 0.8 | Specifies the minimum resource (executor) ratio needed for starting the block distribution. The default value is 0.8, which indicates 80% of the requested resource is allocated for starting block distribution. The minimum value is 0.1 min and the maximum value is 1.0. | -| carbon.search.enabled | false | If set to true, it will use CarbonReader to do distributed scan directly instead of using compute framework like spark, thus avoiding limitation of compute framework like SQL optimizer and task scheduling overhead. | +| carbon.search.enabled (Alpha Feature) | false | If set to true, it will use CarbonReader to do distributed scan directly instead of using compute framework like spark, thus avoiding limitation of compute framework like SQL optimizer and task scheduling overhead. | * **Global Dictionary Configurations** --- End diff -- In Local Dictionary section the following updates needs to be done. 1) Remove the line: â44ad8fb40⦠Updated documentation on Local Dictionary Supoort |â in Page no: 7 at the Local Dictionary Configuration section in the Opensource PDF. 2) Change the description for âLocal dictionary thresholdâ from: âThe maximum cardinality for local dictionary generation (maximum - 100000)â to âThe maximum cardinality for local dictionary generation (maximum value is 100000 and minimum value is 1000. If the âlocal_dictionary_thresholdâ value is set below 1000 or above 100000, then it would take the default value 10000)â --- |
In reply to this post by qiuchenjian-2
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/2603 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7760/ --- |
In reply to this post by qiuchenjian-2
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/2603 Build Failed with Spark 2.2.1, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/6487/ --- |
In reply to this post by qiuchenjian-2
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/2603 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7763/ --- |
In reply to this post by qiuchenjian-2
Github user sgururajshetty commented on a diff in the pull request:
https://github.com/apache/carbondata/pull/2603#discussion_r207516006 --- Diff: docs/configuration-parameters.md --- @@ -140,7 +140,7 @@ This section provides the details of all the configurations required for CarbonD | carbon.enableMinMax | true | Min max is feature added to enhance query performance. To disable this feature, set it false. | | carbon.dynamicallocation.schedulertimeout | 5 | Specifies the maximum time (unit in seconds) the scheduler can wait for executor to be active. Minimum value is 5 sec and maximum value is 15 sec. | | carbon.scheduler.minregisteredresourcesratio | 0.8 | Specifies the minimum resource (executor) ratio needed for starting the block distribution. The default value is 0.8, which indicates 80% of the requested resource is allocated for starting block distribution. The minimum value is 0.1 min and the maximum value is 1.0. | -| carbon.search.enabled | false | If set to true, it will use CarbonReader to do distributed scan directly instead of using compute framework like spark, thus avoiding limitation of compute framework like SQL optimizer and task scheduling overhead. | +| carbon.search.enabled (Alpha Feature) | false | If set to true, it will use CarbonReader to do distributed scan directly instead of using compute framework like spark, thus avoiding limitation of compute framework like SQL optimizer and task scheduling overhead. | * **Global Dictionary Configurations** --- End diff -- The minimum value need not be mentioned now --- |
In reply to this post by qiuchenjian-2
Github user sgururajshetty commented on a diff in the pull request:
https://github.com/apache/carbondata/pull/2603#discussion_r207516087 --- Diff: docs/configuration-parameters.md --- @@ -140,7 +140,7 @@ This section provides the details of all the configurations required for CarbonD | carbon.enableMinMax | true | Min max is feature added to enhance query performance. To disable this feature, set it false. | | carbon.dynamicallocation.schedulertimeout | 5 | Specifies the maximum time (unit in seconds) the scheduler can wait for executor to be active. Minimum value is 5 sec and maximum value is 15 sec. | | carbon.scheduler.minregisteredresourcesratio | 0.8 | Specifies the minimum resource (executor) ratio needed for starting the block distribution. The default value is 0.8, which indicates 80% of the requested resource is allocated for starting block distribution. The minimum value is 0.1 min and the maximum value is 1.0. | -| carbon.search.enabled | false | If set to true, it will use CarbonReader to do distributed scan directly instead of using compute framework like spark, thus avoiding limitation of compute framework like SQL optimizer and task scheduling overhead. | +| carbon.search.enabled (Alpha Feature) | false | If set to true, it will use CarbonReader to do distributed scan directly instead of using compute framework like spark, thus avoiding limitation of compute framework like SQL optimizer and task scheduling overhead. | * **Global Dictionary Configurations** --- End diff -- This issue is handled in a different PR #2576 --- |
In reply to this post by qiuchenjian-2
Github user chetandb commented on the issue:
https://github.com/apache/carbondata/pull/2603 LGTM --- |
In reply to this post by qiuchenjian-2
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/2603 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/7769/ --- |
In reply to this post by qiuchenjian-2
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/2603 Build Success with Spark 2.2.1, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/6493/ --- |
In reply to this post by qiuchenjian-2
Github user kunal642 commented on the issue:
https://github.com/apache/carbondata/pull/2603 LGTM --- |
In reply to this post by qiuchenjian-2
|
In reply to this post by qiuchenjian-2
Github user ravipesala commented on the issue:
https://github.com/apache/carbondata/pull/2603 SDV Build Fail , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/6146/ --- |
Free forum by Nabble | Edit this page |