[jira] [Updated] (CARBONDATA-3815) Insert into table select from another table throws exception for spatial tables

classic Classic list List threaded Threaded
1 message Options
Reply | Threaded
Open this post in threaded view
|

[jira] [Updated] (CARBONDATA-3815) Insert into table select from another table throws exception for spatial tables

Akash R Nilugal (Jira)

     [ https://issues.apache.org/jira/browse/CARBONDATA-3815?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Venugopal Reddy K updated CARBONDATA-3815:
------------------------------------------
    Summary: Insert into table select from another table throws exception for spatial tables  (was: Insert into table select from another table throws exception)

> Insert into table select from another table throws exception for spatial tables
> -------------------------------------------------------------------------------
>
>                 Key: CARBONDATA-3815
>                 URL: https://issues.apache.org/jira/browse/CARBONDATA-3815
>             Project: CarbonData
>          Issue Type: Bug
>          Components: core, spark-integration
>    Affects Versions: 2.0.0
>            Reporter: Venugopal Reddy K
>            Priority: Major
>
> *Issue:*
> Insert into table select from another table throws exception for spatial tables. NoSuchElementException exception is thrown with 'mygeohash' column.
> {color:#FF0000}Exception in thread "main" java.util.NoSuchElementException: key not found: mygeohashException in thread "main" java.util.NoSuchElementException: key not found: mygeohash at scala.collection.MapLike$class.default(MapLike.scala:228) at scala.collection.AbstractMap.default(Map.scala:59) at scala.collection.mutable.HashMap.apply(HashMap.scala:65) at org.apache.spark.sql.execution.command.management.CarbonInsertIntoCommand$$anonfun$getReArrangedIndexAndSelectedSchema$5.apply(CarbonInsertIntoCommand.scala:504) at org.apache.spark.sql.execution.command.management.CarbonInsertIntoCommand$$anonfun$getReArrangedIndexAndSelectedSchema$5.apply(CarbonInsertIntoCommand.scala:497) at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) at org.apache.spark.sql.execution.command.management.CarbonInsertIntoCommand.getReArrangedIndexAndSelectedSchema(CarbonInsertIntoCommand.scala:496) at org.apache.spark.sql.execution.command.management.CarbonInsertIntoCommand.processData(CarbonInsertIntoCommand.scala:164){color}
> *Step to reproduce:*
>  # Create source table and target table spatial tables.
>  # Load data to source table.
>  # Insert into target table select from source table.
> *TestCase:*
> spark.sql(s"""
>  | CREATE TABLE source(
>  | timevalue BIGINT,
>  | longitude LONG,
>  | latitude LONG) COMMENT "This is a GeoTable"
>  | STORED AS carbondata
>  | TBLPROPERTIES ('INDEX_HANDLER'='mygeohash',
>  | 'INDEX_HANDLER.mygeohash.type'='geohash',
>  | 'INDEX_HANDLER.mygeohash.sourcecolumns'='longitude, latitude',
>  | 'INDEX_HANDLER.mygeohash.originLatitude'='39.832277',
>  | 'INDEX_HANDLER.mygeohash.gridSize'='50',
>  | 'INDEX_HANDLER.mygeohash.minLongitude'='115.811865',
>  | 'INDEX_HANDLER.mygeohash.maxLongitude'='116.782233',
>  | 'INDEX_HANDLER.mygeohash.minLatitude'='39.832277',
>  | 'INDEX_HANDLER.mygeohash.maxLatitude'='40.225281',
>  | 'INDEX_HANDLER.mygeohash.conversionRatio'='1000000')
>  """.stripMargin)
> val path = s"$rootPath/examples/spark/src/main/resources/geodata.csv"
> // scalastyle:off
> spark.sql(
>  s"""
>  | LOAD DATA LOCAL INPATH '$path'
>  | INTO TABLE source
>  | OPTIONS('COMPLEX_DELIMITER_LEVEL_1'='#')
>  """.stripMargin)
> spark.sql(s"""
>  | CREATE TABLE target(
>  | timevalue BIGINT,
>  | longitude LONG,
>  | latitude LONG) COMMENT "This is a GeoTable"
>  | STORED AS carbondata
>  | TBLPROPERTIES ('INDEX_HANDLER'='mygeohash',
>  | 'INDEX_HANDLER.mygeohash.type'='geohash',
>  | 'INDEX_HANDLER.mygeohash.sourcecolumns'='longitude, latitude',
>  | 'INDEX_HANDLER.mygeohash.originLatitude'='39.832277',
>  | 'INDEX_HANDLER.mygeohash.gridSize'='50',
>  | 'INDEX_HANDLER.mygeohash.minLongitude'='115.811865',
>  | 'INDEX_HANDLER.mygeohash.maxLongitude'='116.782233',
>  | 'INDEX_HANDLER.mygeohash.minLatitude'='39.832277',
>  | 'INDEX_HANDLER.mygeohash.maxLatitude'='40.225281',
>  | 'INDEX_HANDLER.mygeohash.conversionRatio'='1000000')
>  """.stripMargin)
> spark.sql("insert into target select * from source")



--
This message was sent by Atlassian Jira
(v8.3.4#803005)