Re: [DISCUSSION] Support write Flink streaming data to Carbon

Posted by Jacky Li-3 on
URL: http://apache-carbondata-dev-mailing-list-archive.168.s1.nabble.com/DISCUSSION-Support-write-Flink-streaming-data-to-Carbon-tp85670p85714.html

+1 for these feature, in my opinion, flink-carbon is a good fit for near realtiem analytics

One doubt is that in your design, the Collect Segment command and Compaction command are two separate commands, right?

Collect Segment command modify the metadata files (tablestatus file and segment file), while Compaction command merges small data files and build indexes.

Is my understanding right?

Regards,
Jacky

On 2019/10/29 06:59:51, "爱在西元前" <[hidden email]> wrote:

> The write process is:
>
> Write flink streaming data to local file system of flink task node use flink StreamingFileSink and carbon SDK;
>
> Copy local carbon data file to carbon data store system, such as HDFS, S3;
>
> Generate and write segment file to ${tablePath}/load_details;
>
> Run "alter table ${tableName} collect segments" command on server, to compact segment files in ${tablePath}/load_details, and then move the compacted segment file to ${tablePath}/Metadata/Segments/,update table status file finally.
>
> Have raised a jira https://issues.apache.org/jira/browse/CARBONDATA-3557
>
> Welcome you opinion and suggestions.