[jira] [Created] (FLINK-10126) There should be a Scala DataSource

2018-08-12 Thread Alexis Sarda-Espinosa (JIRA)
Alexis Sarda-Espinosa created FLINK-10126:
-

 Summary: There should be a Scala DataSource
 Key: FLINK-10126
 URL: https://issues.apache.org/jira/browse/FLINK-10126
 Project: Flink
  Issue Type: Improvement
Reporter: Alexis Sarda-Espinosa


In Java, an ExecutionEnvironment's createInput method returns a DataSource, 
whereas the Scala version returns a DataSet. There is no Scala DataSource 
wrapper, and the Scala DataSet does not provide the Java DataSource methods, 
such as getSplitDataProperties.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-10127) Add TypeInformation and serializers for JDK7 Instant

2018-08-12 Thread Alexey Trenikhin (JIRA)
Alexey Trenikhin created FLINK-10127:


 Summary: Add TypeInformation  and serializers for JDK7 Instant
 Key: FLINK-10127
 URL: https://issues.apache.org/jira/browse/FLINK-10127
 Project: Flink
  Issue Type: Improvement
Reporter: Alexey Trenikhin


Currently Flink's basic types include all Java primitives and their boxed form, 
plus {{void}}, {{String}}, {{Date}}, {{BigDecimal}}, and {{BigInteger}}. New 
JDK7 Instance type should be added as well



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-10128) Timeout on transfering jars on travis-ci

2018-08-12 Thread buptljy (JIRA)
buptljy created FLINK-10128:
---

 Summary: Timeout on transfering jars on travis-ci
 Key: FLINK-10128
 URL: https://issues.apache.org/jira/browse/FLINK-10128
 Project: Flink
  Issue Type: Bug
  Components: Build System
Reporter: buptljy


07:55:06.549 [ERROR] Failed to execute goal on project flink-parent: Could not 
resolve dependencies for project 
org.apache.flink:flink-parent:pom:1.7-SNAPSHOT: Could not transfer artifact 
com.google.code.findbugs:jsr305:jar:1.3.9 from/to central 
(http://repo.maven.apache.org/maven2): Connect to repo.maven.apache.org:80 
[repo.maven.apache.org/151.101.184.215] failed: Connection timed out 
(Connection timed out) -> [Help 1]

[Travis Page|https://travis-ci.org/apache/flink/jobs/415047040]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


[jira] [Created] (FLINK-10129) Flink job IDs are not getting deleted automatically from zookeeper metadata after canceling flink job in flink HA cluster

2018-08-12 Thread Keshav Lodhi (JIRA)
Keshav Lodhi created FLINK-10129:


 Summary: Flink job IDs are not getting deleted automatically from 
zookeeper metadata after canceling flink job in flink HA cluster 
 Key: FLINK-10129
 URL: https://issues.apache.org/jira/browse/FLINK-10129
 Project: Flink
  Issue Type: Bug
Reporter: Keshav Lodhi


Hi Team,

Here is, what i am looking for:
 * We have  flink HA dockerized cluster with (3 zookeepers, 2 job-managers, 3 
task-managers) 
 * So whenever we are cancelling the flink job, it is getting cancelled but it 
is not deleting the cancelled job ID from the zookeeper metadata (Inside 
flink/jobgraph folder in zookeeper) automatically. 
 * So whenever any one of the job-manager goes down/restarted , it doesn't come 
up and throws exception like  "Could not find this job id xx".
 * The current work around is to remove the canceled job ID from the zookeeper 
metadata manually. (But this is not the recommended solution).     

 

Please advise.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)


Re: [Discuss] FLIP-26 - SSL Mutual Authentication

2018-08-12 Thread Stephan Ewen
Sounds good, Eron!

Please go ahead...

On Sat, Jul 28, 2018 at 1:33 AM, Eron Wright  wrote:

>  As an update to this thread, Stephan opted to split the internal/external
> configuration (by providing overrides for a common SSL configuration):
> https://github.com/apache/flink/pull/6326
>
> Note that Akka doesn't support hostname verification in its 'classic'
> remoting implementation (though the new Artery implementation apparently
> does), and such verification wouldn't apply to the client certificate
> anyway.   So the reality is that one should use a limited truststore (never
> the system truststore) for Akka communication.
>
> On the question of routing external communication thru the YARN resource
> proxy or Mesos/DCOS admin router, the value proposition is:
> a) simplifies service discovery on the part of external clients,
> b) permits single sign-on (SSO) be delegating authentication to a central
> authority,
> c) facilitates access from outside the cluster, via a public address.
> The main challenge is that the Flink client code must support a more
> diverse array of authentication methods, e.g. Kerberos when communicating
> with the YARN proxy.
>
> Given #6326, the next steps would be (unordered):
> a) create an umbrella issue for the overall effort
> b) dive into the authorization work for external communication
> c) implement auto-generation of a certificate for internal communication
> d) implement TLS on queryable state interface (FLINK-5029)
>
> I'll take care of (a) unless there is any objection.
> -Eron
>
>
> On Sun, May 13, 2018 at 5:45 AM Stephan Ewen 
> wrote:
>
> > Throwing in some more food for thought:
> >
> > An alternative to the above proposed separation of internal and external
> > SSL would be the following:
> >
> >   - We separate channel encryption and authentication
> >   - We use one common SSL layer (internal and external) that is in both
> > cases only responsible for establishing an encrypted connection
> >   - Authentication / authorization internally is done by SASL with
> > username/password or shared secret.
> >   - Authentication externally must be through a proxy and authorization
> > based on a validating HTTP headers set by the proxy, as discussed above..
> >
> > Advantages:
> >   - There is only one certificate needed, which could also be shared
> across
> > applications
> >   - One or two lines in the config authenticate and authorize internal
> > communication
> >   - One could possibly still fall back to the other mode by skipping
> >
> > Open Questions / Disadvantages
> >   - Given that hostname verification during SSL handshake is not possible
> > in many setups, the encrypted channel is vulnerable to man-in-the-middle
> > attacks without mutual authentication. Not sure how serious that is,
> > because it would need an attacker to have compromise network nodes of the
> > cluster already. is that not a universal issue in the K8s world?
> >
> > This is anyways a bit hypothetical, because as long as we have akka
> beneath
> > the RPC layer, we cannot go with that approach.
> >
> > However, if we want to at least keep the door open towards something like
> > that in the future, we would need to set up configuration in such a way
> > that we have a "common SSL" configuration (keystore, truststore, etc.)
> and
> > internal/external options that override those. That would anyways be
> > helpful for backwards compatibility.
> >
> > @Eron - what are your thoughts on that?
> >
> >
> >
> >
> >
> >
> >
> >
> > On Sun, May 13, 2018 at 1:40 AM, Stephan Ewen 
> > wrote:
> >
> > > Thank you for bringing this proposal up. It looks very good and we seem
> > to
> > > be thinking along very similar lines.
> > >
> > > Below are some comments and thoughts on the FLIP.
> > >
> > > *Internal vs. External Connectivity*
> > >
> > > That is a very helpful distinction, let's build on that.
> > >
> > >   - I would suggest to treat eventually all communication coming
> > > potentially from users as external, meaning Client-to-Dispatcher,
> > > Client-to-JobManager (trigger savepoint, change parallelism, ...), Web
> > UI,
> > > Queryable State.
> > >
> > >   - That leaves communication that is only between
> > JobManager/TaskManager/
> > > ResourceManager/Dispatcher/HistoryServer as internal.
> > >
> > >   - I am somewhat operating under the assumption that all external
> > > communication will eventually be HTTP/REST. That works best with many
> > > setups and is the basis for using service proxies that
> > > handle  authentication/authorization.
> > >
> > >
> > > In Flink 1.5 and future versions, we have the following update there:
> > >
> > >   - Akka is now strictly internal connectivity, the client (except
> legacy
> > > client) do not use it any more.
> > >
> > >   - The Blob Server will move to purely internal connectivity in Flink
> > > 1.6, where a POST of a job to the Dispatcher has the jars and the
> > JobGraph.
> > > That is important for Kubernetes setups, where exposing the