Login  Register

Re: carbondata partitioned by date generate many small files

Posted by Jacky Li on Jun 08, 2018; 8:07am
URL: http://apache-carbondata-dev-mailing-list-archive.168.s1.nabble.com/carbondata-partitioned-by-date-generate-many-small-files-tp51475p51644.html

Hi,

I couldn’t see the picture you sent, can you send a text of it?

Regards,
Jacky

> 在 2018年6月6日,上午9:46,陈星宇 <[hidden email]> 写道:
>
> hi Li,
> Yes,i got the partition folder as you say, but under the partition folder ,there are many small file just like following picture,
> How to merge then automatically after jobs done.
>
>
> thanks
>
> ChenXingYu
>  
>  
> ------------------ Original ------------------
> From:  "Jacky Li"<[hidden email]>;
> Date:  Tue, Jun 5, 2018 08:43 PM
> To:  "dev"<[hidden email]>;
> Subject:  Re: carbondata partitioned by date generate many small files
>  
> Hi,
>
> There is a testcase in StandardPartitionTableQueryTestCase used date column as partition column, if you run that testcase, the partition folder generated looks like following picture.
>  
>
> Are you getting similar folders?
>
> Regards,
> Jacky
>
>> 在 2018年6月5日,下午2:49,陈星宇 <[hidden email] <mailto:[hidden email]>> 写道:
>>
>> hi carbondata team,
>>
>>
>> i am using carbondata 1.3.1 to create table and import data, generated many small files and spark job is very slow, i suspected the number of file is related to the number of spark job . but if i decrease the jobs, job will fail because of outofmemory. see my ddl as below:
>>
>>
>> create table xx.xx(
>> dept_name string,
>> xx
>> .
>> .
>> .
>> ) PARTITIONED BY (xxx date)
>> STORED BY 'carbondata' TBLPROPERTIES('SORT_COLUMNS'='xxx,xxx,xxx ,xxx,xxx')
>>
>>
>>
>> please give some advice.
>>
>>
>> thanks
>>
>>
>> ChenXingYu
>