[
https://issues.apache.org/jira/browse/CARBONDATA-267?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15535577#comment-15535577 ]
ASF GitHub Bot commented on CARBONDATA-267:
-------------------------------------------
Github user sujith71955 commented on a diff in the pull request:
https://github.com/apache/incubator-carbondata/pull/189#discussion_r81310297
--- Diff: integration/spark/src/main/scala/org/apache/spark/sql/execution/command/carbonTableSchema.scala ---
@@ -419,6 +420,7 @@ class TableNewProcessor(cm: tableModel, sqlContext: SQLContext) {
schemaEvol
.setSchemaEvolutionEntryList(new util.ArrayList[SchemaEvolutionEntry]())
tableSchema.setTableId(UUID.randomUUID().toString)
+ tableSchema.setBlocksize(Integer.parseInt(cm.tableBlockSize.getOrElse(0).toString))
--- End diff --
What if user will provide value as 1024M or 1MB, you will get exception and the same has not been handled.
I think we need to handle
> Set block_size for table on table level
> ---------------------------------------
>
> Key: CARBONDATA-267
> URL:
https://issues.apache.org/jira/browse/CARBONDATA-267> Project: CarbonData
> Issue Type: New Feature
> Affects Versions: 0.1.0-incubating
> Reporter: zhangshunyu
> Assignee: zhangshunyu
> Fix For: 0.2.0-incubating
>
>
> Set block_size for table on table level
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)