areyouokfreejoe closed pull request #4086: URL: https://github.com/apache/carbondata/pull/4086 ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
areyouokfreejoe commented on pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#issuecomment-771456273 > If we enable the property `ENABLE_AUTO_LOAD_MERGE` then which segment id are we planning to show, the segment generated after compaction or before compaction? Better to add a test case for that scenario also. > If we enable the property `ENABLE_AUTO_LOAD_MERGE` then which segment id are we planning to show, the segment generated after compaction or before compaction? Better to add a test case for that scenario also. If we enable 'AUTO_LOAD_MERGE',then we return and show the segment id before compaction since the user would focus on his load operation. Test case has been added. Please review. ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
areyouokfreejoe commented on pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#issuecomment-771458076 > If we enable the property `ENABLE_AUTO_LOAD_MERGE` then which segment id are we planning to show, the segment generated after compaction or before compaction? Better to add a test case for that scenario also. ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
nihal0107 commented on pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#issuecomment-771464559 > > If we enable the property `ENABLE_AUTO_LOAD_MERGE` then which segment id are we planning to show, the segment generated after compaction or before compaction? Better to add a test case for that scenario also. > > > If we enable the property `ENABLE_AUTO_LOAD_MERGE` then which segment id are we planning to show, the segment generated after compaction or before compaction? Better to add a test case for that scenario also. > > If we enable 'AUTO_LOAD_MERGE',then we return and show the segment id before compaction since the user would focus on his load operation. Test case has been added. Please review. We are showing the segment id because if we need to query on specific segment then this feature will be helpful. But if we will show the segment id before compaction and will query on specific segment then the operation will fail. ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
nihal0107 edited a comment on pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#issuecomment-771464559 > > If we enable the property `ENABLE_AUTO_LOAD_MERGE` then which segment id are we planning to show, the segment generated after compaction or before compaction? Better to add a test case for that scenario also. > > > If we enable the property `ENABLE_AUTO_LOAD_MERGE` then which segment id are we planning to show, the segment generated after compaction or before compaction? Better to add a test case for that scenario also. > > If we enable 'AUTO_LOAD_MERGE',then we return and show the segment id before compaction since the user would focus on his load operation. Test case has been added. Please review. We are showing the segment id because if we need to query on specific segment then this feature will be helpful. But if we will show the segment id before compaction and will query on specific segment then the operation will fail. Better to take the opinion in community. ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
CarbonDataQA2 commented on pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#issuecomment-771515497 Build Failed with Spark 2.4.5, Please check CI http://121.244.95.60:12444/job/ApacheCarbon_PR_Builder_2.4.5/3645/ ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
CarbonDataQA2 commented on pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#issuecomment-771518487 Build Success with Spark 2.3.4, Please check CI http://121.244.95.60:12444/job/ApacheCarbonPRBuilder2.3/5405/ ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
ajantha-bhat commented on pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#issuecomment-771560715 @nihal0107 : As per me, showing the original single segment before compaction is ok, because it is load command and not the compaction command. so, when load command finishes we can give segment id that loaded and user can do show segments before querying that segment to see if it has undergone compaction or not. ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
ajantha-bhat commented on pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#issuecomment-771560807 retest this please ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
ajantha-bhat commented on a change in pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#discussion_r568528970 ########## File path: integration/spark/src/test/scala/org/apache/spark/util/CarbonCommandSuite.scala ########## @@ -82,6 +83,43 @@ class CarbonCommandSuite extends QueryTest with BeforeAndAfterAll { """.stripMargin) } + protected def createTestTable(tableName: String): Unit = { + sql( + s""" Review comment: Instead of adding new testcase , In one of the existing testcases of **loading, insert into, partition table loading, partition table insert into**, just add a validation for segment id ########## File path: integration/spark/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonInsertIntoCommand.scala ########## @@ -276,7 +280,15 @@ case class CarbonInsertIntoCommand(databaseNameOp: Option[String], } throw ex } - Seq.empty + if(loadResultForReturn!=null && loadResultForReturn.getLoadName!=null) { + Seq(Row(loadResultForReturn.getLoadName)) + } else { + rowsForReturn Review comment: why are you returning number of rows instead of segment id here ? when will the code enter here when load is success ? can you add some comment ? ########## File path: integration/spark/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonInsertIntoCommand.scala ########## @@ -276,7 +280,15 @@ case class CarbonInsertIntoCommand(databaseNameOp: Option[String], } throw ex } - Seq.empty + if(loadResultForReturn!=null && loadResultForReturn.getLoadName!=null) { Review comment: I think code is not formatted, we follow space after if and != ########## File path: integration/spark/src/test/scala/org/apache/spark/util/CarbonCommandSuite.scala ########## @@ -100,6 +138,56 @@ class CarbonCommandSuite extends QueryTest with BeforeAndAfterAll { private lazy val location = CarbonProperties.getStorePath() + test("Return segment ID after load and insert") { + val tableName = "test_table" + val inputTableName = "csv_table" + val inputPath = s"$resourcesPath/data_alltypes.csv" + dropTable(tableName) + dropTable(inputTableName) + createAndLoadInputTable(inputTableName, inputPath) + createTestTable(tableName) + checkAnswer(sql( + s""" + | INSERT INTO TABLE $tableName + | SELECT shortField, intField, bigintField, doubleField, stringField, + | from_unixtime(unix_timestamp(timestampField,'yyyy/M/dd')) timestampField, decimalField, + | cast(to_date(from_unixtime(unix_timestamp(dateField,'yyyy/M/dd'))) as date), charField + | FROM $inputTableName + """.stripMargin), Seq(Row("0"))) + checkAnswer(sql( + s"LOAD DATA LOCAL INPATH '$inputPath'" + + s" INTO TABLE $tableName" + + " OPTIONS('FILEHEADER'=" + + "'shortField,intField,bigintField,doubleField,stringField," + + "timestampField,decimalField,dateField,charField')"), Seq(Row("1"))) Review comment: possible to return text like "Successfully loaded to segment id : 1" instead of returning just "1" ########## File path: integration/spark/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonLoadDataCommand.scala ########## @@ -191,7 +196,15 @@ case class CarbonLoadDataCommand(databaseNameOp: Option[String], throw ex } } - Seq.empty + if(loadResultForReturn!=null && loadResultForReturn.getLoadName!=null) { Review comment: @QiangCai, @ydvpankaj99 : why our checkstyle is not catching this format issues ? ########## File path: integration/spark/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonLoadDataCommand.scala ########## @@ -191,7 +196,15 @@ case class CarbonLoadDataCommand(databaseNameOp: Option[String], throw ex } } - Seq.empty + if(loadResultForReturn!=null && loadResultForReturn.getLoadName!=null) { Review comment: @QiangCai, @ydvpankaj99 : why our checkstyle is not catching this format issues ? ########## File path: integration/spark/src/test/scala/org/apache/spark/util/CarbonCommandSuite.scala ########## @@ -100,6 +138,56 @@ class CarbonCommandSuite extends QueryTest with BeforeAndAfterAll { private lazy val location = CarbonProperties.getStorePath() + test("Return segment ID after load and insert") { + val tableName = "test_table" + val inputTableName = "csv_table" + val inputPath = s"$resourcesPath/data_alltypes.csv" + dropTable(tableName) + dropTable(inputTableName) + createAndLoadInputTable(inputTableName, inputPath) + createTestTable(tableName) + checkAnswer(sql( + s""" + | INSERT INTO TABLE $tableName + | SELECT shortField, intField, bigintField, doubleField, stringField, + | from_unixtime(unix_timestamp(timestampField,'yyyy/M/dd')) timestampField, decimalField, + | cast(to_date(from_unixtime(unix_timestamp(dateField,'yyyy/M/dd'))) as date), charField + | FROM $inputTableName + """.stripMargin), Seq(Row("0"))) + checkAnswer(sql( + s"LOAD DATA LOCAL INPATH '$inputPath'" + + s" INTO TABLE $tableName" + + " OPTIONS('FILEHEADER'=" + + "'shortField,intField,bigintField,doubleField,stringField," + + "timestampField,decimalField,dateField,charField')"), Seq(Row("1"))) + + } + + test("Return segment ID after load and insert when auto merge enabled") { Review comment: same comment as above ########## File path: integration/spark/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonInsertIntoCommand.scala ########## @@ -276,7 +280,15 @@ case class CarbonInsertIntoCommand(databaseNameOp: Option[String], } throw ex } - Seq.empty + if(loadResultForReturn!=null && loadResultForReturn.getLoadName!=null) { + Seq(Row(loadResultForReturn.getLoadName)) + } else { + rowsForReturn Review comment: partition table case, what will be rowsForReturn? ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
ajantha-bhat commented on a change in pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#discussion_r568524277 ########## File path: integration/spark/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonLoadDataCommand.scala ########## @@ -191,7 +196,15 @@ case class CarbonLoadDataCommand(databaseNameOp: Option[String], throw ex } } - Seq.empty + if(loadResultForReturn!=null && loadResultForReturn.getLoadName!=null) { Review comment: @QiangCai, @ydvpankaj99 : why our checkstyle is not catching this format issues ? ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
CarbonDataQA2 commented on pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#issuecomment-771609869 Build Success with Spark 2.4.5, Please check CI http://121.244.95.60:12444/job/ApacheCarbon_PR_Builder_2.4.5/3649/ ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
CarbonDataQA2 commented on pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#issuecomment-771616575 Build Success with Spark 2.3.4, Please check CI http://121.244.95.60:12444/job/ApacheCarbonPRBuilder2.3/5409/ ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
nihal0107 commented on pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#issuecomment-771385539 ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
CarbonDataQA2 commented on pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#issuecomment-771318129 ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
areyouokfreejoe commented on pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#issuecomment-771456273 ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
areyouokfreejoe closed pull request #4086: URL: https://github.com/apache/carbondata/pull/4086 ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
ajantha-bhat commented on pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#issuecomment-771560715 ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
ajantha-bhat commented on a change in pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#discussion_r568528970 ########## File path: integration/spark/src/test/scala/org/apache/spark/util/CarbonCommandSuite.scala ########## @@ -82,6 +83,43 @@ class CarbonCommandSuite extends QueryTest with BeforeAndAfterAll { """.stripMargin) } + protected def createTestTable(tableName: String): Unit = { + sql( + s""" Review comment: Instead of adding new testcase , In one of the existing testcases of **loading, insert into, partition table loading, partition table insert into**, just add a validation for segment id ########## File path: integration/spark/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonInsertIntoCommand.scala ########## @@ -276,7 +280,15 @@ case class CarbonInsertIntoCommand(databaseNameOp: Option[String], } throw ex } - Seq.empty + if(loadResultForReturn!=null && loadResultForReturn.getLoadName!=null) { + Seq(Row(loadResultForReturn.getLoadName)) + } else { + rowsForReturn Review comment: why are you returning number of rows instead of segment id here ? when will the code enter here when load is success ? can you add some comment ? ########## File path: integration/spark/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonInsertIntoCommand.scala ########## @@ -276,7 +280,15 @@ case class CarbonInsertIntoCommand(databaseNameOp: Option[String], } throw ex } - Seq.empty + if(loadResultForReturn!=null && loadResultForReturn.getLoadName!=null) { Review comment: I think code is not formatted, we follow space after if and != ########## File path: integration/spark/src/test/scala/org/apache/spark/util/CarbonCommandSuite.scala ########## @@ -100,6 +138,56 @@ class CarbonCommandSuite extends QueryTest with BeforeAndAfterAll { private lazy val location = CarbonProperties.getStorePath() + test("Return segment ID after load and insert") { + val tableName = "test_table" + val inputTableName = "csv_table" + val inputPath = s"$resourcesPath/data_alltypes.csv" + dropTable(tableName) + dropTable(inputTableName) + createAndLoadInputTable(inputTableName, inputPath) + createTestTable(tableName) + checkAnswer(sql( + s""" + | INSERT INTO TABLE $tableName + | SELECT shortField, intField, bigintField, doubleField, stringField, + | from_unixtime(unix_timestamp(timestampField,'yyyy/M/dd')) timestampField, decimalField, + | cast(to_date(from_unixtime(unix_timestamp(dateField,'yyyy/M/dd'))) as date), charField + | FROM $inputTableName + """.stripMargin), Seq(Row("0"))) + checkAnswer(sql( + s"LOAD DATA LOCAL INPATH '$inputPath'" + + s" INTO TABLE $tableName" + + " OPTIONS('FILEHEADER'=" + + "'shortField,intField,bigintField,doubleField,stringField," + + "timestampField,decimalField,dateField,charField')"), Seq(Row("1"))) Review comment: possible to return text like "Successfully loaded to segment id : 1" instead of returning just "1" ########## File path: integration/spark/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonLoadDataCommand.scala ########## @@ -191,7 +196,15 @@ case class CarbonLoadDataCommand(databaseNameOp: Option[String], throw ex } } - Seq.empty + if(loadResultForReturn!=null && loadResultForReturn.getLoadName!=null) { Review comment: @QiangCai, @ydvpankaj99 : why our checkstyle is not catching this format issues ? ########## File path: integration/spark/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonLoadDataCommand.scala ########## @@ -191,7 +196,15 @@ case class CarbonLoadDataCommand(databaseNameOp: Option[String], throw ex } } - Seq.empty + if(loadResultForReturn!=null && loadResultForReturn.getLoadName!=null) { Review comment: @QiangCai, @ydvpankaj99 : why our checkstyle is not catching this format issues ? ########## File path: integration/spark/src/test/scala/org/apache/spark/util/CarbonCommandSuite.scala ########## @@ -100,6 +138,56 @@ class CarbonCommandSuite extends QueryTest with BeforeAndAfterAll { private lazy val location = CarbonProperties.getStorePath() + test("Return segment ID after load and insert") { + val tableName = "test_table" + val inputTableName = "csv_table" + val inputPath = s"$resourcesPath/data_alltypes.csv" + dropTable(tableName) + dropTable(inputTableName) + createAndLoadInputTable(inputTableName, inputPath) + createTestTable(tableName) + checkAnswer(sql( + s""" + | INSERT INTO TABLE $tableName + | SELECT shortField, intField, bigintField, doubleField, stringField, + | from_unixtime(unix_timestamp(timestampField,'yyyy/M/dd')) timestampField, decimalField, + | cast(to_date(from_unixtime(unix_timestamp(dateField,'yyyy/M/dd'))) as date), charField + | FROM $inputTableName + """.stripMargin), Seq(Row("0"))) + checkAnswer(sql( + s"LOAD DATA LOCAL INPATH '$inputPath'" + + s" INTO TABLE $tableName" + + " OPTIONS('FILEHEADER'=" + + "'shortField,intField,bigintField,doubleField,stringField," + + "timestampField,decimalField,dateField,charField')"), Seq(Row("1"))) + + } + + test("Return segment ID after load and insert when auto merge enabled") { Review comment: same comment as above ########## File path: integration/spark/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonInsertIntoCommand.scala ########## @@ -276,7 +280,15 @@ case class CarbonInsertIntoCommand(databaseNameOp: Option[String], } throw ex } - Seq.empty + if(loadResultForReturn!=null && loadResultForReturn.getLoadName!=null) { + Seq(Row(loadResultForReturn.getLoadName)) + } else { + rowsForReturn Review comment: partition table case, what will be rowsForReturn? ########## File path: integration/spark/src/main/scala/org/apache/spark/sql/execution/command/management/CarbonLoadDataCommand.scala ########## @@ -191,7 +196,15 @@ case class CarbonLoadDataCommand(databaseNameOp: Option[String], throw ex } } - Seq.empty + if(loadResultForReturn!=null && loadResultForReturn.getLoadName!=null) { Review comment: @QiangCai, @ydvpankaj99 : why our checkstyle is not catching this format issues ? ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
In reply to this post by GitBox
nihal0107 edited a comment on pull request #4086: URL: https://github.com/apache/carbondata/pull/4086#issuecomment-771464559 > > If we enable the property `ENABLE_AUTO_LOAD_MERGE` then which segment id are we planning to show, the segment generated after compaction or before compaction? Better to add a test case for that scenario also. > > > If we enable the property `ENABLE_AUTO_LOAD_MERGE` then which segment id are we planning to show, the segment generated after compaction or before compaction? Better to add a test case for that scenario also. > > If we enable 'AUTO_LOAD_MERGE',then we return and show the segment id before compaction since the user would focus on his load operation. Test case has been added. Please review. We are showing the segment id because if we need to query on specific segment then this feature will be helpful. But if we will show the segment id before compaction and will query on specific segment then the operation will fail. Better to take the opinion in community. ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [hidden email] |
Free forum by Nabble | Edit this page |