[ https://issues.apache.org/jira/browse/CARBONDATA-1048?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] anubhav tarar closed CARBONDATA-1048. ------------------------------------- Resolution: Fixed > Update Hive Guide > ------------------ > > Key: CARBONDATA-1048 > URL: https://issues.apache.org/jira/browse/CARBONDATA-1048 > Project: CarbonData > Issue Type: Improvement > Components: hive-integration > Affects Versions: 1.0.0-incubating > Environment: hive 1.2.1,spark2.1 > Reporter: anubhav tarar > Assignee: anubhav tarar > Priority: Trivial > Time Spent: 1h 40m > Remaining Estimate: 0h > > Decimal data type raises exception while selecting the data from the table in hive with steps given in hive guide > 1) In Spark Shell : > a) Create Table - > import org.apache.spark.sql.SparkSession > import org.apache.spark.sql.CarbonSession._ > val carbon = SparkSession.builder().config(sc.getConf).getOrCreateCarbonSession("hdfs://localhost:54310/opt/data") > scala> carbon.sql(""" create table testHive1(id int,name string,scale decimal(10,0),country string,salary double) stored by'carbondata' """).show > b) Load Data - > scala> carbon.sql(""" load data inpath 'hdfs://localhost:54310/Files/testHive1.csv' into table testHive1 """ ).show > 2) In Hive : > a) Add Jars - > add jar /home/neha/incubator-carbondata/assembly/target/scala-2.11/carbondata_2.11-1.1.0-incubating-SNAPSHOT-shade-hadoop2.7.2.jar; > add jar /opt/spark-2.1.0-bin-hadoop2.7/jars/spark-catalyst_2.11-2.1.0.jar; > add jar /home/neha/incubator-carbondata/integration/hive/carbondata-hive-1.1.0-incubating-SNAPSHOT.jar; > b) Create Table - > create table testHive1(id int,name string,scale decimal(10,0),country string,salary double); > c) Alter location - > hive> alter table testHive1 set LOCATION 'hdfs://localhost:54310/opt/data/default/testhive1' ; > d) Set Properties - > set hive.mapred.supports.subdirectories=true; > set mapreduce.input.fileinputformat.input.dir.recursive=true; > d) Alter FileFormat - > alter table testHive1 set FILEFORMAT > INPUTFORMAT "org.apache.carbondata.hive.MapredCarbonInputFormat" > OUTPUTFORMAT "org.apache.carbondata.hive.MapredCarbonOutputFormat" > SERDE "org.apache.carbondata.hive.CarbonHiveSerDe"; > hive> ADD JAR /home/hduser/spark-2.1.0-bin-hadoop2.7/jars/spark-catalyst_2.11-2.1.0.jar; > Added [/home/hduser/spark-2.1.0-bin-hadoop2.7/jars/spark-catalyst_2.11-2.1.0.jar] to class path > Added resources: [/home/hduser/spark-2.1.0-bin-hadoop2.7/jars/spark-catalyst_2.11-2.1.0.jar] > f) Execute Queries - > select * from testHive1; > 3) Query : > hive> select * from testHive1; > Exception in thread "[main][partitionID:hive25;queryID:4537623368167]" java.lang.NoClassDefFoundError: scala/math/Ordered > at java.lang.ClassLoader.defineClass1(Native Method) > at java.lang.ClassLoader.defineClass(ClassLoader.java:763) > at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) > when i add the scala-library and scala-reflect jar it works fine > hive> ADD JAR /home/knoldus/Videos/scala-library-2.11.1.jar; > Added [/home/knoldus/Videos/scala-library-2.11.1.jar] to class path > Added resources: [/home/knoldus/Videos/scala-library-2.11.1.jar] > hive> ADD JAR /home/knoldus/Videos/scala-reflect-2.11.1.jar; > Added [/home/knoldus/Videos/scala-reflect-2.11.1.jar] to class path > Added resources: [/home/knoldus/Videos/scala-reflect-2.11.1.jar] > fired the query again > hive> select * from testHive1; > OK > 2 runlin 2 china 33000.2 > 1 yuhai 2 china 33000.1 > so its better to mention about adding these jar in hive -- This message was sent by Atlassian JIRA (v6.4.14#64029) |
Free forum by Nabble | Edit this page |