GitHub user kevinjmh opened a pull request:
https://github.com/apache/carbondata/pull/1787 [CARBONDATA-2017] Fix input path checking when loading data from multiple paths Be sure to do all of the following checklist to help us incorporate your contribution quickly and easily: - [ ] Any interfaces changed? - [ ] Any backward compatibility impacted? - [ ] Document update required? - [ ] Testing done Please provide details on - Whether new unit test cases have been added or why no new tests are required? - How it is tested? Please attach test report. - Is it a performance related change? Please attach the performance test report. - Any additional information to help reviewers in testing this change. - [ ] For large changes, please consider breaking it into sub-tasks under an umbrella JIRA. You can merge this pull request into a Git repository by running: $ git pull https://github.com/kevinjmh/carbondata load_multi_path Alternatively you can review and apply these changes as the patch at: https://github.com/apache/carbondata/pull/1787.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #1787 ---- ---- --- |
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/1787 Build Failed with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/2681/ --- |
In reply to this post by qiuchenjian-2
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/1787 Build Success with Spark 2.2.1, Please check CI http://88.99.58.216:8080/job/ApacheCarbonPRBuilder/1447/ --- |
In reply to this post by qiuchenjian-2
Github user ravipesala commented on the issue:
https://github.com/apache/carbondata/pull/1787 SDV Build Success , Please check CI http://144.76.159.231:8080/job/ApacheSDVTests/2818/ --- |
In reply to this post by qiuchenjian-2
Github user jackylk commented on a diff in the pull request:
https://github.com/apache/carbondata/pull/1787#discussion_r160845775 --- Diff: integration/spark-common/src/main/scala/org/apache/spark/util/FileUtils.scala --- @@ -73,7 +73,8 @@ object FileUtils { val stringBuild = new StringBuilder() val filePaths = inputPath.split(",") for (i <- 0 until filePaths.size) { - val fileType = FileFactory.getFileType(filePaths(i)) + val filePath = CarbonUtil.checkAndAppendHDFSUrl(filePaths(i)) --- End diff -- This is only for HDFS, right? How about other storage system support like S3? @SangeetaGulia Can you have a look at this, I think this may impact #1584 that you are working on --- |
In reply to this post by qiuchenjian-2
Github user SangeetaGulia commented on a diff in the pull request:
https://github.com/apache/carbondata/pull/1787#discussion_r160882178 --- Diff: integration/spark-common/src/main/scala/org/apache/spark/util/FileUtils.scala --- @@ -73,7 +73,8 @@ object FileUtils { val stringBuild = new StringBuilder() val filePaths = inputPath.split(",") for (i <- 0 until filePaths.size) { - val fileType = FileFactory.getFileType(filePaths(i)) + val filePath = CarbonUtil.checkAndAppendHDFSUrl(filePaths(i)) --- End diff -- @jackylk I have verified this. It is working fine with S3 also. We will now be able to use the carbon property **carbon.ddl.base.hdfs.url** for s3 also to provide base URL. --- |
In reply to this post by qiuchenjian-2
Github user jackylk commented on a diff in the pull request:
https://github.com/apache/carbondata/pull/1787#discussion_r161232844 --- Diff: integration/spark-common/src/main/scala/org/apache/spark/util/FileUtils.scala --- @@ -73,7 +73,8 @@ object FileUtils { val stringBuild = new StringBuilder() val filePaths = inputPath.split(",") for (i <- 0 until filePaths.size) { - val fileType = FileFactory.getFileType(filePaths(i)) + val filePath = CarbonUtil.checkAndAppendHDFSUrl(filePaths(i)) --- End diff -- ok --- |
In reply to this post by qiuchenjian-2
|
In reply to this post by qiuchenjian-2
|
Free forum by Nabble | Edit this page |