[
https://issues.apache.org/jira/browse/FLINK-4316?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15409384#comment-15409384
]
Stephan Ewen commented on FLINK-4316:
-------------------------------------
An alternative would be to copy the Hadoop {{Writable}} type (simple interface)
into {{link-core}}. We did that before, it works well and is a much easier
solution (no reflection work), but comes with two issues:
- We need to carry that class in the flink codebase
- There will be multiple versions of {{Writable}} and in theory, these could
lead to class cast exceptions.
> Make flink-core independent of Hadoop
> -------------------------------------
>
> Key: FLINK-4316
> URL: https://issues.apache.org/jira/browse/FLINK-4316
> Project: Flink
> Issue Type: Bug
> Components: Core
> Affects Versions: 1.1.0
> Reporter: Stephan Ewen
> Assignee: Stephan Ewen
> Fix For: 1.2.0
>
>
> We want to gradually reduce the hard and heavy mandatory dependencies in
> Hadoop. Hadoop will still be part of (most) flink downloads, but the API
> projects should not have a hard dependency on Hadoop.
> I suggest to start with {{flink-core}}, because it only depends on Hadoop for
> the {{Writable}} type, to support seamless operation of Hadoop types.
> I propose to move all {{WritableTypeInfo}}-related classes to the
> {{flink-hadoop-compatibility}} project and access them via reflection in the
> {{TypeExtractor}}.
> That way, {{Writable}} types will be out of the box supported if users have
> the {{flink-hadoop-compatibility}} project in the classpath.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)