Hi everyone, thanks to Robert, I found the problem.
I was setting "recovery.zookeeper.path.root" on the command line with -yD. Apparently this is currently not supported. You need to set it the parameter in flink-conf.yaml. Cheers, Konstantin On 05.04.2016 12:52, Konstantin Knauf wrote: > Hi Robert, > > I tried several paths and rmr before. > > It stopped after 1-2 minutes. There was an exception on the shell. > Sorry, should have attached to the last mail. > > Thanks, > > Konstnatin > > On 05.04.2016 11:22, Robert Metzger wrote: >> I've tried reproducing the issue on a test cluster, but everything >> worked fine. >> >> Have you tried different values for "recovery.zookeeper.path.root" or >> only one? Maybe the path you've put contains invalid data? >> >> Regarding the client log you've send: Did you manually stop the client >> or did it stop after a few minutes? >> The JobManager stops after a few minutes because the client requested a >> shutdown. Usually, the client only shuts down on an exception or when >> the user stops the yarn session. >> There is no exception in the client log. Was there an exception printed >> to the shell? >> >> This log message: >> >> 2016-04-05 08:48:34,912 DEBUG org.apache.flink.yarn.FlinkYarnCluster >> - Received message option None >> >> Should not be an issue. >> >> >> On Tue, Apr 5, 2016 at 10:14 AM, Ufuk Celebi <u...@apache.org >> <mailto:u...@apache.org>> wrote: >> >> Hey Konstantin, >> >> just looked at the logs and the cluster is started, but the job is >> indeed never submitted. >> >> I've forwarded this to Robert, because he is familiar with the YARN >> client. I will look into how the client interacts with the ZooKeeper >> root path. >> >> – Ufuk >> >> >> On Tue, Apr 5, 2016 at 9:18 AM, Konstantin Knauf >> <konstantin.kn...@tngtech.com <mailto:konstantin.kn...@tngtech.com>> >> wrote: >> > Hi Ufuk, Hi Stephan, >> > >> > sorry for the late response Attached the client logs. >> > >> > Cheers, >> > >> > Konstantin >> > >> > On 04.04.2016 21 <tel:04.04.2016%2021>:20, Stephan Ewen wrote: >> >> This seems to the the critical part in the logs: >> >> >> >> 2016-03-31 09:01:52,234 INFO org.apache.flink.yarn.YarnJobManager >> >> - Re-submitting 0 job graphs. >> >> 2016-03-31 09:02:51,182 INFO org.apache.flink.yarn.YarnJobManager >> >> - Stopping YARN JobManager with status FAILED and >> >> diagnostic Flink YARN Client requested shutdown. >> >> >> >> The YarnJobManager starts up properly, but the Client never sends >> >> anything, shuts down at some point, and tears down the YARN cluster. >> >> >> >> Client logs would help a lot there... >> >> >> >> >> >> >> >> >> >> On Sat, Apr 2, 2016 at 12:43 PM, Ufuk Celebi <u...@apache.org >> <mailto:u...@apache.org> >> >> <mailto:u...@apache.org <mailto:u...@apache.org>>> wrote: >> >> >> >> Hey Konstantin, >> >> >> >> That's weird. Can you please log the client output on DEBUG >> level and >> >> provide that as well? I'm wondering whether the client uses a >> >> different root path. >> >> >> >> The following seems to happen: >> >> - you use ledf_recovery as the root namespace >> >> - the task managers are connecting (they resolve the JM >> address via >> >> ZooKeeper in this case as well, which means they correctly >> use the >> >> same namespace) >> >> - but the client, which started the YARN session, does not >> ever submit >> >> the job to the cluster. >> >> >> >> – Ufuk >> >> >> >> On Thu, Mar 31, 2016 at 9:23 AM, Konstantin Knauf >> >> <konstantin.kn...@tngtech.com >> <mailto:konstantin.kn...@tngtech.com> >> <mailto:konstantin.kn...@tngtech.com >> <mailto:konstantin.kn...@tngtech.com>>> >> >> wrote: >> >> > Hi everyone, >> >> > >> >> > we are running in some problems with multiple per-job yarn >> >> sessions, too. >> >> > >> >> > When we are are starting a per-job yarn session (Flink 1.0, >> Hadoop >> >> 2.4) >> >> > with recovery.zookeeper.path.root other than /flink, the >> yarn session >> >> > starts but no job is submitted, and after 1 min or so the >> session >> >> > crashes. I attached the jobmanager log. >> >> > >> >> > In Zookeeper the root-directory is created and child-nodes >> >> > >> >> > leaderlatch >> >> > jobgraphs >> >> > >> >> > /flink does also exist, but does not have child nodes. >> >> > >> >> > Everything runs fine, with the default >> recovery.zookeeper.root.path. >> >> > >> >> > Does anyone have an idea, what is going on? >> >> > >> >> > Cheers, >> >> > >> >> > Konstnatin >> >> > >> >> > >> >> > On 23.11.2015 17:00, Gwenhael Pasquiers wrote: >> >> >> We are not yet using HA in our cluster instances. >> >> >> >> >> >> But yes, we will have to change the zookeeper.path.root J >> >> >> >> >> >> >> >> >> >> >> >> We package our jobs with their own config folder (we don’t >> rely on >> >> >> flink’s config folder); we can put the maven project name >> into this >> >> >> property then they will have different values J >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> *From:*Till Rohrmann [mailto:trohrm...@apache.org >> <mailto:trohrm...@apache.org> >> >> <mailto:trohrm...@apache.org <mailto:trohrm...@apache.org>>] >> >> >> *Sent:* lundi 23 novembre 2015 14:51 >> >> >> *To:* user@flink.apache.org <mailto:user@flink.apache.org> >> <mailto:user@flink.apache.org <mailto:user@flink.apache.org>> >> >> >> *Subject:* Re: YARN High Availability >> >> >> >> >> >> >> >> >> >> >> >> The problem is the execution graph handle which is stored in >> >> ZooKeeper. >> >> >> You can manually remove it via the ZooKeeper shell by >> simply deleting >> >> >> everything below your `recovery.zookeeper.path.root` >> ZNode. But you >> >> >> should be sure that the cluster has been stopped before. >> >> >> >> >> >> >> >> >> >> >> >> Do you start the different clusters with different >> >> >> `recovery.zookeeper.path.root` values? If not, then you should >> >> run into >> >> >> troubles when running multiple clusters at the same time. The >> >> reason is >> >> >> that then all clusters will think that they belong together. >> >> >> >> >> >> >> >> >> >> >> >> Cheers, >> >> >> >> >> >> Till >> >> >> >> >> >> >> >> >> >> >> >> On Mon, Nov 23, 2015 at 2:15 PM, Gwenhael Pasquiers >> >> >> <gwenhael.pasqui...@ericsson.com >> <mailto:gwenhael.pasqui...@ericsson.com> >> >> <mailto:gwenhael.pasqui...@ericsson.com >> <mailto:gwenhael.pasqui...@ericsson.com>> >> >> >> <mailto:gwenhael.pasqui...@ericsson.com >> <mailto:gwenhael.pasqui...@ericsson.com> >> >> <mailto:gwenhael.pasqui...@ericsson.com >> <mailto:gwenhael.pasqui...@ericsson.com>>>> wrote: >> >> >> >> >> >> OK, I understand. >> >> >> >> >> >> Maybe we are not really using flink as you intended. The >> way we are >> >> >> using it, one cluster equals one job. That way we are sure >> to isolate >> >> >> the different jobs as much as possible and in case of >> crashes / >> >> bugs / >> >> >> (etc) can completely kill one cluster without interfering with >> >> the other >> >> >> jobs. >> >> >> >> >> >> That future behavior seems good :-) >> >> >> >> >> >> Instead of the manual flink commands, is there to manually >> delete >> >> those >> >> >> old jobs before launching my job ? They probably are >> somewhere in >> >> hdfs, >> >> >> aren't they ? >> >> >> >> >> >> B.R. >> >> >> >> >> >> >> >> >> >> >> >> -----Original Message----- >> >> >> From: Ufuk Celebi [mailto:u...@apache.org >> <mailto:u...@apache.org> <mailto:u...@apache.org >> <mailto:u...@apache.org>> >> >> <mailto:u...@apache.org <mailto:u...@apache.org> >> <mailto:u...@apache.org <mailto:u...@apache.org>>>] >> >> >> Sent: lundi 23 novembre 2015 12:12 >> >> >> To: user@flink.apache.org <mailto:user@flink.apache.org> >> <mailto:user@flink.apache.org <mailto:user@flink.apache.org>> >> >> <mailto:user@flink.apache.org <mailto:user@flink.apache.org> >> <mailto:user@flink.apache.org <mailto:user@flink.apache.org>>> >> >> >> Subject: Re: YARN High Availability >> >> >> >> >> >> Hey Gwenhaël, >> >> >> >> >> >> the restarting jobs are most likely old job submissions. >> They are not >> >> >> cleaned up when you shut down the cluster, but only when >> they finish >> >> >> (either regular finish or after cancelling). >> >> >> >> >> >> The workaround is to use the command line frontend: >> >> >> >> >> >> bin/flink cancel JOBID >> >> >> >> >> >> for each RESTARTING job. Sorry about the inconvenience! >> >> >> >> >> >> We are in an active discussion about addressing this. The >> future >> >> >> behaviour will be that the startup or shutdown of a >> cluster cleans up >> >> >> everything and an option to skip this step. >> >> >> >> >> >> The reasoning for the initial solution (not removing >> anything) was to >> >> >> make sure that no jobs are deleted by accident. But it >> looks like >> >> this >> >> >> is more confusing than helpful. >> >> >> >> >> >> – Ufuk >> >> >> >> >> >>> On 23 Nov 2015, at 11:45, Gwenhael Pasquiers >> >> >> <gwenhael.pasqui...@ericsson.com >> <mailto:gwenhael.pasqui...@ericsson.com> >> >> <mailto:gwenhael.pasqui...@ericsson.com >> <mailto:gwenhael.pasqui...@ericsson.com>> >> >> >> <mailto:gwenhael.pasqui...@ericsson.com >> <mailto:gwenhael.pasqui...@ericsson.com> >> >> <mailto:gwenhael.pasqui...@ericsson.com >> <mailto:gwenhael.pasqui...@ericsson.com>>>> wrote: >> >> >>> >> >> >>> Hi again ! >> >> >>> >> >> >>> On the same topic I'm still trying to start my streaming job >> >> with HA. >> >> >>> The HA part seems to be more or less OK (I killed the >> JobManager and >> >> >> it came back), however I have an issue with the TaskManagers. >> >> >>> I configured my job to have only one TaskManager and 1 >> slot that >> >> does >> >> >> [source=>map=>sink]. >> >> >>> The issue I'm encountering is that other instances of my >> job appear >> >> >> and are in the RESTARTING status since there is only one >> task slot. >> >> >>> >> >> >>> Do you know of this, or have an idea of where to look in >> order to >> >> >> understand what's happening ? >> >> >>> >> >> >>> B.R. >> >> >>> >> >> >>> Gwenhaël PASQUIERS >> >> >>> >> >> >>> -----Original Message----- >> >> >>> From: Maximilian Michels [mailto:m...@apache.org >> <mailto:m...@apache.org> >> >> <mailto:m...@apache.org <mailto:m...@apache.org>> >> <mailto:m...@apache.org <mailto:m...@apache.org> >> <mailto:m...@apache.org <mailto:m...@apache.org>>>] >> >> >>> Sent: jeudi 19 novembre 2015 13:36 >> >> >>> To: user@flink.apache.org <mailto:user@flink.apache.org> >> <mailto:user@flink.apache.org <mailto:user@flink.apache.org>> >> >> <mailto:user@flink.apache.org <mailto:user@flink.apache.org> >> <mailto:user@flink.apache.org <mailto:user@flink.apache.org>>> >> >> >>> Subject: Re: YARN High Availability >> >> >>> >> >> >>> The docs have been updated. >> >> >>> >> >> >>> On Thu, Nov 19, 2015 at 12:36 PM, Ufuk Celebi >> <u...@apache.org <mailto:u...@apache.org> >> >> <mailto:u...@apache.org <mailto:u...@apache.org>> >> >> >> <mailto:u...@apache.org <mailto:u...@apache.org> >> <mailto:u...@apache.org <mailto:u...@apache.org>>>> wrote: >> >> >>>> I’ve added a note about this to the docs and asked Max >> to trigger a >> >> >> new build of them. >> >> >>>> >> >> >>>> Regarding Aljoscha’s idea: I like it. It is essentially >> a shortcut >> >> >> for configuring the root path. >> >> >>>> >> >> >>>> In any case, it is orthogonal to Till’s proposals. That >> one we need >> >> >> to address as well (see FLINK-2929). The motivation for >> the current >> >> >> behaviour was to be rather defensive when removing state >> in order >> >> to not >> >> >> loose data accidentally. But it can be confusing, indeed. >> >> >>>> >> >> >>>> – Ufuk >> >> >>>> >> >> >>>>> On 19 Nov 2015, at 12:08, Till Rohrmann >> <trohrm...@apache.org <mailto:trohrm...@apache.org> >> >> <mailto:trohrm...@apache.org <mailto:trohrm...@apache.org>> >> >> >> <mailto:trohrm...@apache.org <mailto:trohrm...@apache.org> >> <mailto:trohrm...@apache.org <mailto:trohrm...@apache.org>>>> wrote: >> >> >>>>> >> >> >>>>> You mean an additional start-up parameter for the >> >> `start-cluster.sh` >> >> >> script for the HA case? That could work. >> >> >>>>> >> >> >>>>> On Thu, Nov 19, 2015 at 11:54 AM, Aljoscha Krettek >> >> >> <aljos...@apache.org <mailto:aljos...@apache.org> >> <mailto:aljos...@apache.org <mailto:aljos...@apache.org>> >> >> <mailto:aljos...@apache.org <mailto:aljos...@apache.org> >> <mailto:aljos...@apache.org <mailto:aljos...@apache.org>>>> wrote: >> >> >>>>> Maybe we could add a user parameter to specify a >> cluster name that >> >> >> is used to make the paths unique. >> >> >>>>> >> >> >>>>> >> >> >>>>> On Thu, Nov 19, 2015, 11:24 Till Rohrmann >> >> <trohrm...@apache.org <mailto:trohrm...@apache.org> >> <mailto:trohrm...@apache.org <mailto:trohrm...@apache.org>> >> >> >> <mailto:trohrm...@apache.org <mailto:trohrm...@apache.org> >> <mailto:trohrm...@apache.org <mailto:trohrm...@apache.org>>>> wrote: >> >> >>>>> I agree that this would make the configuration easier. >> However, it >> >> >> entails also that the user has to retrieve the randomized path >> >> from the >> >> >> logs if he wants to restart jobs after the cluster has >> crashed or >> >> >> intentionally restarted. Furthermore, the system won't be >> able to >> >> clean >> >> >> up old checkpoint and job handles in case that the cluster >> stop was >> >> >> intentional. >> >> >>>>> >> >> >>>>> Thus, the question is how do we define the behaviour in >> order to >> >> >> retrieve handles and to clean up old handles so that ZooKeeper >> >> won't be >> >> >> cluttered with old handles? >> >> >>>>> >> >> >>>>> There are basically two modes: >> >> >>>>> >> >> >>>>> 1. Keep state handles when shutting down the cluster. >> Provide >> >> a mean >> >> >> to define a fixed path when starting the cluster and also >> a mean to >> >> >> purge old state handles. Furthermore, add a shutdown mode >> where the >> >> >> handles under the current path are directly removed. This >> mode would >> >> >> guarantee to always have the state handles available if not >> >> explicitly >> >> >> told differently. However, the downside is that ZooKeeper >> will be >> >> >> cluttered most certainly. >> >> >>>>> >> >> >>>>> 2. Remove the state handles when shutting down the cluster. >> >> Provide >> >> >> a shutdown mode where we keep the state handles. This will >> keep >> >> >> ZooKeeper clean but will give you also the possibility to >> keep a >> >> >> checkpoint around if necessary. However, the user is more >> likely >> >> to lose >> >> >> his state when shutting down the cluster. >> >> >>>>> >> >> >>>>> On Thu, Nov 19, 2015 at 10:55 AM, Robert Metzger >> >> >> <rmetz...@apache.org <mailto:rmetz...@apache.org> >> <mailto:rmetz...@apache.org <mailto:rmetz...@apache.org>> >> >> <mailto:rmetz...@apache.org <mailto:rmetz...@apache.org> >> <mailto:rmetz...@apache.org <mailto:rmetz...@apache.org>>>> wrote: >> >> >>>>> I agree with Aljoscha. Many companies install Flink >> (and its >> >> config) >> >> >> in a central directory and users share that installation. >> >> >>>>> >> >> >>>>> On Thu, Nov 19, 2015 at 10:45 AM, Aljoscha Krettek >> >> >> <aljos...@apache.org <mailto:aljos...@apache.org> >> <mailto:aljos...@apache.org <mailto:aljos...@apache.org>> >> >> <mailto:aljos...@apache.org <mailto:aljos...@apache.org> >> <mailto:aljos...@apache.org <mailto:aljos...@apache.org>>>> wrote: >> >> >>>>> I think we should find a way to randomize the paths >> where the HA >> >> >> stuff stores data. If users don’t realize that they store >> data in the >> >> >> same paths this could lead to problems. >> >> >>>>> >> >> >>>>>> On 19 Nov 2015, at 08:50, Till Rohrmann >> <trohrm...@apache.org <mailto:trohrm...@apache.org> >> >> <mailto:trohrm...@apache.org <mailto:trohrm...@apache.org>> >> >> >> <mailto:trohrm...@apache.org <mailto:trohrm...@apache.org> >> <mailto:trohrm...@apache.org <mailto:trohrm...@apache.org>>>> wrote: >> >> >>>>>> >> >> >>>>>> Hi Gwenhaël, >> >> >>>>>> >> >> >>>>>> good to hear that you could resolve the problem. >> >> >>>>>> >> >> >>>>>> When you run multiple HA flink jobs in the same >> cluster, then you >> >> >> don’t have to adjust the configuration of Flink. It should >> work >> >> out of >> >> >> the box. >> >> >>>>>> >> >> >>>>>> However, if you run multiple HA Flink cluster, then >> you have >> >> to set >> >> >> for each cluster a distinct ZooKeeper root path via the option >> >> >> recovery.zookeeper.path.root in the Flink configuraiton. >> This is >> >> >> necessary because otherwise all JobManagers (the ones of the >> >> different >> >> >> clusters) will compete for a single leadership. >> Furthermore, all >> >> >> TaskManagers will only see the one and only leader and >> connect to it. >> >> >> The reason is that the TaskManagers will look up their >> leader at >> >> a ZNode >> >> >> below the ZooKeeper root path. >> >> >>>>>> >> >> >>>>>> If you have other questions then don’t hesitate asking me. >> >> >>>>>> >> >> >>>>>> Cheers, >> >> >>>>>> Till >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> On Wed, Nov 18, 2015 at 6:37 PM, Gwenhael Pasquiers >> >> >> <gwenhael.pasqui...@ericsson.com >> <mailto:gwenhael.pasqui...@ericsson.com> >> >> <mailto:gwenhael.pasqui...@ericsson.com >> <mailto:gwenhael.pasqui...@ericsson.com>> >> >> >> <mailto:gwenhael.pasqui...@ericsson.com >> <mailto:gwenhael.pasqui...@ericsson.com> >> >> <mailto:gwenhael.pasqui...@ericsson.com >> <mailto:gwenhael.pasqui...@ericsson.com>>>> wrote: >> >> >>>>>> Nevermind, >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> Looking at the logs I saw that it was having issues >> trying to >> >> >> connect to ZK. >> >> >>>>>> >> >> >>>>>> To make I short is had the wrong port. >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> It is now starting. >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> Tomorrow I’ll try to kill some JobManagers *evil*. >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> Another question : if I have multiple HA flink jobs, are >> >> there some >> >> >> points to check in order to be sure that they won’t collide on >> >> hdfs or ZK ? >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> B.R. >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> Gwenhaël PASQUIERS >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> From: Till Rohrmann [mailto:till.rohrm...@gmail.com >> <mailto:till.rohrm...@gmail.com> >> >> <mailto:till.rohrm...@gmail.com <mailto:till.rohrm...@gmail.com>> >> >> >> <mailto:till.rohrm...@gmail.com >> <mailto:till.rohrm...@gmail.com> <mailto:till.rohrm...@gmail.com >> <mailto:till.rohrm...@gmail.com>>>] >> >> >>>>>> Sent: mercredi 18 novembre 2015 18:01 >> >> >>>>>> To: user@flink.apache.org >> <mailto:user@flink.apache.org> <mailto:user@flink.apache.org >> <mailto:user@flink.apache.org>> >> >> <mailto:user@flink.apache.org <mailto:user@flink.apache.org> >> <mailto:user@flink.apache.org <mailto:user@flink.apache.org>>> >> >> >>>>>> Subject: Re: YARN High Availability >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> Hi Gwenhaël, >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> do you have access to the yarn logs? >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> Cheers, >> >> >>>>>> >> >> >>>>>> Till >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> On Wed, Nov 18, 2015 at 5:55 PM, Gwenhael Pasquiers >> >> >> <gwenhael.pasqui...@ericsson.com >> <mailto:gwenhael.pasqui...@ericsson.com> >> >> <mailto:gwenhael.pasqui...@ericsson.com >> <mailto:gwenhael.pasqui...@ericsson.com>> >> >> >> <mailto:gwenhael.pasqui...@ericsson.com >> <mailto:gwenhael.pasqui...@ericsson.com> >> >> <mailto:gwenhael.pasqui...@ericsson.com >> <mailto:gwenhael.pasqui...@ericsson.com>>>> wrote: >> >> >>>>>> >> >> >>>>>> Hello, >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> We’re trying to set up high availability using an existing >> >> >> zookeeper quorum already running in our Cloudera cluster. >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> So, as per the doc we’ve changed the max attempt in >> yarn’s config >> >> >> as well as the flink.yaml. >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> recovery.mode: zookeeper >> >> >>>>>> >> >> >>>>>> recovery.zookeeper.quorum: >> host1:3181,host2:3181,host3:3181 >> >> >>>>>> >> >> >>>>>> state.backend: filesystem >> >> >>>>>> >> >> >>>>>> state.backend.fs.checkpointdir: hdfs:///flink/checkpoints >> >> >>>>>> >> >> >>>>>> recovery.zookeeper.storageDir: hdfs:///flink/recovery/ >> >> >>>>>> >> >> >>>>>> yarn.application-attempts: 1000 >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> Everything is ok as long as recovery.mode is commented. >> >> >>>>>> >> >> >>>>>> As soon as I uncomment recovery.mode the deployment on >> yarn is >> >> >> stuck on : >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> “Deploying cluster, current state ACCEPTED”. >> >> >>>>>> >> >> >>>>>> “Deployment took more than 60 seconds….” >> >> >>>>>> >> >> >>>>>> Every second. >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> And I have more than enough resources available on my yarn >> >> cluster. >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> Do you have any idea of what could cause this, and/or >> what logs I >> >> >> should look for in order to understand ? >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> B.R. >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> Gwenhaël PASQUIERS >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>>> >> >> >>>>> >> >> >>>>> >> >> >>>>> >> >> >>>>> >> >> >>>> >> >> >>> <unwanted_jobs.jpg> >> >> >> >> >> >> >> >> >> >> >> > >> >> > -- >> >> > Konstantin Knauf * konstantin.kn...@tngtech.com >> <mailto:konstantin.kn...@tngtech.com> >> >> <mailto:konstantin.kn...@tngtech.com >> <mailto:konstantin.kn...@tngtech.com>> * +49-174-3413182 >> <tel:%2B49-174-3413182> >> >> <tel:%2B49-174-3413182> >> >> > TNG Technology Consulting GmbH, Betastr. 13a, 85774 >> Unterföhring >> >> > Geschäftsführer: Henrik Klagges, Christoph Stock, Dr. >> Robert Dahlke >> >> > Sitz: Unterföhring * Amtsgericht München * HRB 135082 >> >> >> >> >> > >> > -- >> > Konstantin Knauf * konstantin.kn...@tngtech.com >> <mailto:konstantin.kn...@tngtech.com> * +49-174-3413182 >> <tel:%2B49-174-3413182> >> > TNG Technology Consulting GmbH, Betastr. 13a, 85774 Unterföhring >> > Geschäftsführer: Henrik Klagges, Christoph Stock, Dr. Robert Dahlke >> > Sitz: Unterföhring * Amtsgericht München * HRB 135082 >> >> > -- Konstantin Knauf * konstantin.kn...@tngtech.com * +49-174-3413182 TNG Technology Consulting GmbH, Betastr. 13a, 85774 Unterföhring Geschäftsführer: Henrik Klagges, Christoph Stock, Dr. Robert Dahlke Sitz: Unterföhring * Amtsgericht München * HRB 135082