[ https://issues.apache.org/jira/browse/FLINK-10929?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16816192#comment-16816192 ]
Stephan Ewen commented on FLINK-10929: -------------------------------------- There are two different considerations: (1) Arrow for interoperability with other systems and languages (source / sink / inter-process-communication) (2) Arrow as a format for internal data processing in SQL / Table API. For (1), I see it would make sense, but then we need to look more concretely at what we want to integrate. Arrow is not a magic integration, it is a data format. For (2), the original Flink query processor is not getting much committer attention, because we plan to replace it with the Blink processor in the long run. The Blink processor is in the process or merging and that needs to be finished before we can start making more changes. [~ykt836] could probably provide some input into when would be a good time to follow up there. > Add support for Apache Arrow > ---------------------------- > > Key: FLINK-10929 > URL: https://issues.apache.org/jira/browse/FLINK-10929 > Project: Flink > Issue Type: Wish > Components: Runtime / State Backends > Reporter: Pedro Cardoso Silva > Priority: Minor > Attachments: image-2019-04-10-13-43-08-107.png > > > Investigate the possibility of adding support for Apache Arrow as a > standardized columnar, memory format for data. > Given the activity that [https://github.com/apache/arrow] is currently > getting and its claims objective of providing a zero-copy, standardized data > format across platforms, I think it makes sense for Flink to look into > supporting it. -- This message was sent by Atlassian JIRA (v7.6.3#76005)