[
https://issues.apache.org/jira/browse/CARBONDATA-308?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15644343#comment-15644343 ]
ASF GitHub Bot commented on CARBONDATA-308:
-------------------------------------------
Github user jackylk commented on a diff in the pull request:
https://github.com/apache/incubator-carbondata/pull/262#discussion_r86786673
--- Diff: hadoop/src/main/java/org/apache/carbondata/hadoop/CarbonInputFormat.java ---
@@ -311,80 +278,6 @@ private void addSegmentsIfEmpty(JobContext job, AbsoluteTableIdentifier absolute
return result;
}
- /**
- * get total number of rows. Same as count(*)
- *
- * @throws IOException
- * @throws IndexBuilderException
- */
- public long getRowCount(JobContext job) throws IOException, IndexBuilderException {
--- End diff --
ok
> Use CarbonInputFormat in CarbonScanRDD compute
> ----------------------------------------------
>
> Key: CARBONDATA-308
> URL:
https://issues.apache.org/jira/browse/CARBONDATA-308> Project: CarbonData
> Issue Type: Sub-task
> Components: spark-integration
> Reporter: Jacky Li
> Assignee: Jacky Li
> Fix For: 0.2.0-incubating
>
>
> Take CarbonScanRDD as the target RDD, modify as following:
> 1. In driver side, only getSplit is required, so only filter condition is required, no need to create full QueryModel object, so we can move creation of QueryModel from driver side to executor side.
> 2. use CarbonInputFormat.createRecordReader in CarbonScanRDD.compute instead of use QueryExecutor directly
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)