Login  Register

Re: FileNotFoundExceptions while running CarbonData

Posted by Swapnil Shinde on Jul 17, 2017; 10:49pm
URL: http://apache-carbondata-dev-mailing-list-archive.168.s1.nabble.com/FileNotFoundExceptions-while-running-CarbonData-tp18340p18368.html

Thanks. I think I fixed it support maprFS. I will do some more testing and
then add a jira ticket and PR.

On Mon, Jul 17, 2017 at 11:51 AM, Ravindra Pesala <[hidden email]>
wrote:

> Hi,
>
> Right now we don't have support to maprfs filesystem , so it would be
> unpredictable even though you have fixed at some places. We need to check
> in all places and add the maprfs support. So it would be great if you can
> add the support to maprfs in carbon.
>
> And one more observation is please provide absolute path along with
> maprfs:// in all the places instead of giving relative path. And also make
> sure that storelocation inside carbon properties and store location while
> creating carbon session must be same.
>
> Regards,
> Ravindra.
>
> On 17 July 2017 at 11:25, Swapnil Shinde <[hidden email]> wrote:
>
> > Hello
> >     I am new to carbon data and we are trying to use carbon data in
> > production. I built and installed it on Spark edge nodes as per given
> > instruction -
> >
> > *Build -* No major issues.
> > *Installation -* Followed yarn installation (
> http://carbondata.apache.org/
> > installation-guide.html)  instructions.
> > *Infrastructure -* Spark 2.1.0 on MapR cluster.
> > *carbon.properties changes -*
> >           carbon.storelocation=/tmp/hacluster/Opt/CarbonStore
> >           carbon.badRecords.location=/opt/Carbon/Spark/badrecords
> >           carbon.lock.type=HDFSLOCK
> > *spark-default.conf changes -*
> >          spark.yarn.dist.files     /opt/mapr/spark/spark-2.1.0/
> > conf/carbon.properties
> >          spark.yarn.dist.archives    /opt/mapr/spark/spark-2.1.0/
> > carbonlib/carbondata.tar.gz
> >          spark.executor.extraJavaOptions
> -Dcarbon.properties.filepath=
> > carbon.properties
> >          spark.driver.extraJavaOptions     -Dcarbon.properties.filepath=/
> > opt/mapr/spark/spark-2.1.0/conf/carbon.properties
> > *Command line -*
> > /opt/mapr/spark/spark-2.1.0/bin/spark-shell --name "My app" --master
> yarn
> > --jars /opt/mapr/spark/spark-2.1.0/carbonlib/carbondata_2.11-1.1.
> 0-shade-hadoop2.2.0.jar
> > \
> > --driver-memory 1g \
> > --executor-cores 2 \
> > --executor-memory 2G
> > *Code snippet -*
> > import org.apache.spark.sql.SparkSession
> > import org.apache.spark.sql.CarbonSession._
> > val carbon = SparkSession.builder().config(sc.getConf).
> > getOrCreateCarbonSession("/mapr/ri0.abc.com/tmp", "/mapr/ri0.abc.com/tmp
> ")
> > carbon.sql("""CREATE TABLE
> >                         IF NOT EXISTS test_table(
> >                                   id string,
> >                                   name string,
> >                                   city string,
> >                                   age Int)
> >                        STORED BY 'carbondata'""")
> >
> > carbon.sql("""LOAD DATA INPATH '/mapr/ri0.comscore.com/tmp/sample.csv'
> >                   INTO TABLE test_table""")
> >
> > First error -
> >       Inital error was "*Dictionay file is locked for updation*". Further
> > debugging showed that it was due to missing maprFS filesyste.
> > (HDFSFileLock.java line # 52) -
> > String hdfsPath =  "conf.get(CarbonCommonConstants.FS_DEFAULT_FS)"
> >
> > I added some code to workaround with path like maprfs:///* and that
> seemed
> > to be working fine. (like adding MAPRFS FileType)
> >
> > *Second error -*
> >    First error was gone after mapFS refactoring but then it fails with
> > below error. *It seems *.dict & *.dictmeta are not getting created.*
> > Could you please help me resolving this error?
> > [image: Inline image 1]
> >
> >
> > Thanks
> > Swapnil
> >
> >
>
>
> --
> Thanks & Regards,
> Ravi
>