I just stopped the cluster with stop-cluster.sh but I had to manually kill the root process because it was not able to terminate it using the aforementioned script. Then I restarted the cluster via start-cluster.sh and now all processes run with the user it was supposed to. Probably once in the past I started the services with sudo and then I was convinced to restart the cluster using the start/stop scripts but the job manager was never restarted actually.. However I didn't get any error about that, I was just reading
"No jobmanager daemon (pid: XXXX) is running anymore on myhost.test.it" Maybe the scripts could be improved to check such a situation? Thanks for the support, Flavio On Mon, Oct 26, 2015 at 3:14 PM, Flavio Pompermaier <pomperma...@okkam.it> wrote: > Yes, the job manager starts as a root process, while taskmanagers with my > user..is that normal? > I was convinced that start-cluster.sh was starting all processes with the > same user :O > > On Mon, Oct 26, 2015 at 3:09 PM, Maximilian Michels <m...@apache.org> > wrote: > >> Hi Flavio, >> >> Are you runing your Flink cluster with root permissions? The directory to >> hold the output splits are created by the JobManager. So if you run then >> JobManager with root permissions, it will create a folder owned by root. If >> the task managers are not run with root permissions, this could be a >> problem. >> >> Cheers, >> Max >> >> On Mon, Oct 26, 2015 at 2:40 PM, Flavio Pompermaier <pomperma...@okkam.it >> > wrote: >> >>> Hi to all, >>> when I run my job within my hadoop cluster (both from command line and >>> from webapp) the output of my job (HDFS) works fine until I set the write >>> parallelism to 1 (the output file is created with the user running the >>> job). If I leave the default parallelism (>1) the job fails because it >>> creates a folder where the owner of the output folder is the root user and >>> the job cannot write the files of my user in that folder anymore. Am I >>> doing something wrong? >>> >>> >>> Best, >>> Flavio >>> >>> >> >