Hi Loris,
Am 29.03.2019 um 14:01 schrieb Loris Bennett:
Hi Marcus,
Marcus Wagner <wag...@itc.rwth-aachen.de> writes:
Hi Loris,
On 3/25/19 1:42 PM, Loris Bennett wrote:
3. salloc works fine too without --x11, subsequent srun with a x11 app works
great
Doing 'salloc' followed by 'ssh -X' works for us too, which is surprising to
me.
This last option currently seems to me to be the best option for users,
being slightly less confusing than logging into the login node again
from the login node, which is our current workaround.
Still, it's all a bit odd.
I assume, you use pam_slurm_adopt?
Yes.
Then it is clear, that this is working and has nothing to do with the x11
forwarding feature of slurm. This is plain ssh X11-forwarding in this case.
OK, I see that, but if I don't need --x11 with salloc, what is it
for? Just to control to control on which nodes forwarding is done
viz. --x11[=<all|first|last>]? What might be a use-case for not having
X11 forwarding for all the nodes, which is the default?
The default is (according to the manpage) 'batch', which means the node,
where the batchscript will be executed (the first of the allocation, I
think).
I do not know what first or last should be intended.
In fact I do not have a use case for x11-forwarding to all nodes, might
have to think a little bit more about that one.
Please keep in mind, that processes started with an adopted ssh session are in
the jobs cgroup (good), but are accounted in the 'extern' step of the job.
e.g.
* sbatch --wrap "sleep 10m"
* ssh to compute-node
* do some work in the compute node
after job is done
* sacct -j <jobid> -o JobID,JobName,MaxRSS,CPUTime,TotalCPU
JobID JobName MaxRSS CPUTime TotalCPU
------------ ---------- ---------- ---------- ----------
1053837 wrap 00:01:42 02:00.159
1053837.bat+ batch 412K 00:01:43 00:00.158
1053837.ext+ extern 543880K 00:01:42 02:00.001
That's interesting, although is there any advantage/difference compared
with just doing
srun --x11 --pty bash
?
With
srun --x11 --pty bash
the accounting will be in the batch step of the job, that is the only
difference I'm aware of at the moment.
With LSF we used that kind of mechanism to start e.g. vtune directly out
of the job. Without the X11-Forwarding feature of Slurm you would have
to salloc some hosts and then ssh to the nodes with x11 forwarding
enabled to then start vtune.
So it is a little bit more to do for the user if you do not do
X11-Forwarding the SLURM style.
Best
Marcus
Cheers,
Loris