Re: [slurm-users] salloc with bash scripts problem

2019-01-03 Thread Renfro, Michael
Those errors appear to pop up when qemu can’t find enough RAM to run. If the #SBATCH lines are only applicable for ‘sbatch' and not ‘srun' or ‘salloc', the ‘--mem=8G' setting there doesn’t affect anything. - Does the srun version of the command work if you specify 'qemu-system-x86_64 -m 2048' o

Re: [slurm-users] Reservation to exceed time limit on a partition for a user

2019-01-03 Thread Carsten Beyer
Hi Matthew, we use QOS for this and add it to the SLURM user, who needs to exceed the partition time limit. You can also set a time limit in the QOS, so that a user cannot exceed the 'limits' too much. Example from our system with 8 hour runlimit per job: # grep -i qos slurm.conf PriorityWei

Re: [slurm-users] How to delete an association

2019-01-03 Thread Fulcomer, Samuel
Great. Yes, I forgot to mention that running or pending jobs can prevent deletion of this information. This makes scripting/automating all the sacctmgr functions somewhat difficult. regards, Sam On Thu, Jan 3, 2019 at 10:18 AM Jianwen Wei wrote: > Thank you, Samuel. I've successfully delete th

Re: [slurm-users] Reservation to exceed time limit on a partition for a user

2019-01-03 Thread Matthew BETTINGER
Answering my own question here. I created a hidden partition which shows like this PartitionName=FOO AllowGroups=ALL AllowAccounts=rt AllowQos=ALL AllocNodes=ALL Default=NO QoS=N/A DefaultTime=NONE DisableRootJobs=NO ExclusiveUser=YES GraceTime=0 Hidden=YES MaxNodes=UNLIMITED MaxTim

Re: [slurm-users] How to delete an association

2019-01-03 Thread Jianwen Wei
Thank you, Samuel. I've successfully delete the association with the following command after the users' jobs completes. # sacctmgr delete user where name=clschf partition=k80 account=acct-clschf Best, Jianwen > On Dec 29, 2018, at 11:50, Fulcomer, Samuel wrote: > > .

[slurm-users] Reservation to exceed time limit on a partition for a user

2019-01-03 Thread Matthew BETTINGER
Hello, We are running slurm 17.02.6 with accounting on a cray CLE system. We currently have a 24 hour job run limit on our partitions and a user needs to run a job which will exceed 24 hours runtime. I tried to make a reservation as seen below allocating the user 36 hours to run his job but it

Re: [slurm-users] Visualisation -- Slurm and (Turbo)VNC

2019-01-03 Thread John Hearns
Hi David. I set up DCV on a cluster of workstations at a facility not far from you a few years ago (in Woking...). I'm not sure what the relevance of having multiple GPUs is - I thought the DCV documentation dealt with that ?? One thing you should do is introduce MobaXterm to your users if they

Re: [slurm-users] Visualisation -- Slurm and (Turbo)VNC

2019-01-03 Thread Daniel Letai
I haven't done this in a long time, but this blog entry might be of some use (I believe I did something similar when required in the past) : https://summerofhpc.prace-ri.eu/remote-accelerated-graphics-with-virtualgl-and-turbovnc/ On 03/01/2019 12:14:52, Bake

[slurm-users] Visualisation -- Slurm and (Turbo)VNC

2019-01-03 Thread Baker D . J .
Hello, We have set up our NICE/DCV cluster and that is proving to be very popular. There are, however, users who would benefit from using the resources offered by our nodes with multiple GPU cards. This potentially means setting up TurboVNC, for example. I would, if possible, like to be able t

Re: [slurm-users] salloc with bash scripts problem

2019-01-03 Thread Mahmood Naderan
Mark Hahn, Using srun only returns a memory allocation error while salloc doesn't [mahmood@rocks7 ~]$ srun --spankx11 ./run_qemu.sh qemu-system-x86_64: -usbdevice tablet: '-usbdevice' is deprecated, please use '-device usb-...' instead qemu-system-x86_64: warning: host doesn't support requested