|
Hi,
I got a strange error that carbonContext can not read parquet file on hdfs
in my program. I test on spark shell and it returns the same error,
anything goes wrong?
>>>>>>Test for sqlContext, Success
----------------------------------------------------------------------------------------------------------------
Spark context available as sc (master = yarn-client, app id =
application_1479961381214_0551).
SQL context available as sqlContext.
scala> val parquetFile = sqlContext.read.parquet("/
user/hive/default/testdata_parquet_all")
parquetFile: org.apache.spark.sql.DataFrame = [id: double,***************...
>>>>>>Test for CarbonContext, Failed
------------------------------------------------------------------------------------------------------------------
Spark context available as sc (master = yarn-client, app id =
application_1479961381214_0552).Hi
SQL context available as sqlContext.
scala> import org.apache.spark.sql.CarbonContext
import org.apache.spark.sql.CarbonContext
scala> val cc = new CarbonContext(sc)
cc: org.apache.spark.sql.CarbonContext = org.apache.spark.sql.
CarbonContext@3574122f
scala> val parquetFile = cc.read.parquet("/user/hive/
default/testdata_parquet_all")
AUDIT 30-11 13:42:24,114 - [*******][appuser][Thread-1]Creating timestamp
file for .
java.io.IOException: No such file or directory
at java.io.UnixFileSystem.createFileExclusively(Native Method)
at java.io.File.createNewFile(File.java:1006)
at org.apache.carbondata.core.datastorage.store.impl.
FileFactory.createNewFile(FileFactory.java:372)
at org.apache.spark.sql.hive.CarbonMetastoreCatalog.
updateSchemasUpdatedTime(CarbonMetastoreCatalog.scala:468)
at org.apache.spark.sql.hive.CarbonMetastoreCatalog.loadMetadata(
CarbonMetastoreCatalog.scala:181)
at org.apache.spark.sql.hive.CarbonMetastoreCatalog.<init>(
CarbonMetastoreCatalog.scala:114)
|