jackylk commented on a change in pull request #3475: [WIP] Optimize getSplits() during compaction
URL: https://github.com/apache/carbondata/pull/3475#discussion_r349925039 ########## File path: integration/spark-common/src/main/scala/org/apache/carbondata/spark/rdd/CarbonMergerRDD.scala ########## @@ -369,44 +371,47 @@ class CarbonMergerRDD[K, V]( } } } + validSegIds.add(eachSeg.getSegmentNo) + } - // map for keeping the relation of a task and its blocks. - job.getConfiguration.set(CarbonTableInputFormat.INPUT_SEGMENT_NUMBERS, eachSeg.getSegmentNo) + // map for keeping the relation of a task and its blocks. + job.getConfiguration + .set(CarbonTableInputFormat.INPUT_SEGMENT_NUMBERS, validSegIds.asScala.mkString(",")) - if (updateStatusManager.getUpdateStatusDetails.length != 0) { - updateDetails = updateStatusManager.getInvalidTimestampRange(eachSeg.getSegmentNo) - } - val updated: Boolean = updateStatusManager.getUpdateStatusDetails.length != 0 - // get splits - val splits = format.getSplits(job) + val updated: Boolean = updateStatusManager.getUpdateStatusDetails.length != 0 + // get splits + val splits = format.getSplits(job) - // keep on assigning till last one is reached. - if (null != splits && splits.size > 0) { - splitsOfLastSegment = splits.asScala - .map(_.asInstanceOf[CarbonInputSplit]) - .filter { split => FileFormat.COLUMNAR_V3.equals(split.getFileFormat) }.toList.asJava - } - val filteredSplits = splits.asScala.map(_.asInstanceOf[CarbonInputSplit]).filter{ entry => - val blockInfo = new TableBlockInfo(entry.getFilePath, - entry.getStart, entry.getSegmentId, - entry.getLocations, entry.getLength, entry.getVersion, - updateStatusManager.getDeleteDeltaFilePath( - entry.getFilePath, - Segment.toSegment(entry.getSegmentId).getSegmentNo) - ) - (!updated || (updated && (!CarbonUtil - .isInvalidTableBlock(blockInfo.getSegmentId, blockInfo.getFilePath, - updateDetails, updateStatusManager)))) && - FileFormat.COLUMNAR_V3.equals(entry.getFileFormat) - } - if (rangeColumn != null) { - totalTaskCount = totalTaskCount + - CarbonCompactionUtil.getTaskCountForSegment(filteredSplits.toArray) + // keep on assigning till last one is reached. + if (null != splits && splits.size > 0) { + splitsOfLastSegment = splits.asScala + .map(_.asInstanceOf[CarbonInputSplit]) + .filter { split => FileFormat.COLUMNAR_V3.equals(split.getFileFormat) }.toList.asJava + } + val filteredSplits = splits.asScala.map(_.asInstanceOf[CarbonInputSplit]).filter { entry => + val segmentId = Segment.toSegment(entry.getSegmentId).getSegmentNo + val blockInfo = new TableBlockInfo(entry.getFilePath, + entry.getStart, entry.getSegmentId, + entry.getLocations, entry.getLength, entry.getVersion, + updateStatusManager.getDeleteDeltaFilePath( + entry.getFilePath, + segmentId) + ) + if (updateStatusManager.getUpdateStatusDetails.length != 0) { + updateDetails = updateStatusManager.getInvalidTimestampRange(segmentId) } - carbonInputSplits ++:= filteredSplits - allSplits.addAll(filteredSplits.asJava) + (!updated || (updated && (!CarbonUtil + .isInvalidTableBlock(blockInfo.getSegmentId, blockInfo.getFilePath, + updateDetails, updateStatusManager)))) && + FileFormat.COLUMNAR_V3.equals(entry.getFileFormat) + } + if (rangeColumn != null) { + totalTaskCount = totalTaskCount + + CarbonCompactionUtil.getTaskCountForSegment(filteredSplits.toArray) } + carbonInputSplits ++:= filteredSplits Review comment: use `ArrayBuffer` instead of `Seq` ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] With regards, Apache Git Services |
Free forum by Nabble | Edit this page |