GitHub user QiangCai opened a pull request:
https://github.com/apache/carbondata/pull/1530 [CARBONDATA-1766] fix serialization issue for CarbonAppendableStreamSink fix serialization issue for CarbonAppendableStreamSink. - [ x] Any interfaces changed? no - [x] Any backward compatibility impacted? no - [x] Document update required? no - [ x] Testing done Please provide details on - Whether new unit test cases have been added or why no new tests are required? - How it is tested? Please attach test report. - Is it a performance related change? Please attach the performance test report. - Any additional information to help reviewers in testing this change. - [ x] For large changes, please consider breaking it into sub-tasks under an umbrella JIRA. You can merge this pull request into a Git repository by running: $ git pull https://github.com/QiangCai/carbondata test Alternatively you can review and apply these changes as the patch at: https://github.com/apache/carbondata/pull/1530.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #1530 ---- commit 2c5605573782b4aca8d0b21bad8edf200028b730 Author: QiangCai <[hidden email]> Date: 2017-11-18T07:27:54Z fix serialization issue of CarbonLoadModel ---- --- |
Github user CarbonDataQA commented on the issue:
https://github.com/apache/carbondata/pull/1530 Build Success with Spark 2.1.0, Please check CI http://136.243.101.176:8080/job/ApacheCarbonPRBuilder1/1272/ --- |
In reply to this post by qiuchenjian-2
|
In reply to this post by qiuchenjian-2
Github user jackylk commented on a diff in the pull request:
https://github.com/apache/carbondata/pull/1530#discussion_r151840214 --- Diff: hadoop/src/main/java/org/apache/carbondata/hadoop/streaming/CarbonStreamRecordWriter.java --- @@ -99,13 +100,21 @@ initialize(job); } + public CarbonStreamRecordWriter(TaskAttemptContext job, CarbonLoadModel carbonLoadModel) + throws IOException { + this.carbonLoadModel = carbonLoadModel; + initialize(job); + } + private void initialize(TaskAttemptContext job) throws IOException { // set basic information hadoopConf = job.getConfiguration(); - CarbonLoadModel carbonLoadModel = CarbonStreamOutputFormat.getCarbonLoadModel(hadoopConf); if (carbonLoadModel == null) { - throw new IOException( - "CarbonStreamRecordWriter require configuration: mapreduce.output.carbon.load.model"); + carbonLoadModel = CarbonStreamOutputFormat.getCarbonLoadModel(hadoopConf); + if (carbonLoadModel == null) { + throw new IOException( + "CarbonStreamRecordWriter require configuration: mapreduce.output.carbon.load.model"); --- End diff -- This configuration should start with `carbon`, change to `carbon.load.model` --- |
In reply to this post by qiuchenjian-2
Github user jackylk commented on a diff in the pull request:
https://github.com/apache/carbondata/pull/1530#discussion_r151840239 --- Diff: hadoop/src/main/java/org/apache/carbondata/hadoop/streaming/CarbonStreamRecordWriter.java --- @@ -99,13 +100,21 @@ initialize(job); } + public CarbonStreamRecordWriter(TaskAttemptContext job, CarbonLoadModel carbonLoadModel) + throws IOException { + this.carbonLoadModel = carbonLoadModel; + initialize(job); + } + private void initialize(TaskAttemptContext job) throws IOException { // set basic information hadoopConf = job.getConfiguration(); - CarbonLoadModel carbonLoadModel = CarbonStreamOutputFormat.getCarbonLoadModel(hadoopConf); if (carbonLoadModel == null) { - throw new IOException( - "CarbonStreamRecordWriter require configuration: mapreduce.output.carbon.load.model"); + carbonLoadModel = CarbonStreamOutputFormat.getCarbonLoadModel(hadoopConf); + if (carbonLoadModel == null) { + throw new IOException( + "CarbonStreamRecordWriter require configuration: mapreduce.output.carbon.load.model"); --- End diff -- Do you really require the whole CarbonLoadModel? It seems you only need CarbonTable object, right? --- |
In reply to this post by qiuchenjian-2
|
In reply to this post by qiuchenjian-2
|
Free forum by Nabble | Edit this page |