nevermind... after reading my email, i just realized the typo. On Thu, Jun 21, 2018 at 11:22 AM Anson Abraham <anson.abra...@gmail.com> wrote:
> Thanks, > However, i'm on ubuntu, so when i installed w/ apt install slurm-llnl and > slurm-llnl-basic-plugins{-dev} > > i added the con_res entry. however, the controller can't find the > select/con_res plugin. Is there something else i need to install to get > the plugin to be recognized? > > On Wed, Jun 20, 2018 at 2:49 AM Thomas HAMEL <hm...@t-hamel.fr> wrote: > >> Hello, >> >> Slurm can have a behavior similar to the one you were used on SGE, it's >> not the default but it's really commonly used. This is called Consumable >> Resources in Slurm, you don't allocate full nodes but resources on the >> node like cores or memory. It's activated in slurm.conf (after >> restarting or reconfiguring the controller) : >> >> SelectType=select/cons_res >> SelectTypeParameters=CR_Core_Memory >> >> https://slurm.schedmd.com/cons_res.html >> >> MPI serves another purpose that do not really fit you model. The >> solution to your issue is Consumable Resources but you can take a look >> at Job Arrays to group all jobs of your loop together. >> >> Regards, >> >> Thomas HAMEL >> >> >> On 19/06/2018 23:33, Anson Abraham wrote: >> > Hi, >> > Relatively new to Slurm. I've been using Sun GridEngine mostly. >> > I have a cluster of 3 machines each with 8 cores. In SGE i allocate >> > the PE slots per machine, where if i submit 24 jobs it'll run all 24 >> > jobs (b/c each job will use 1 core). >> > however, if i submit job in Slurm through sbatch i can only get it to >> > run 3 jobs at a time, even when i define the cpus_per_task. I was >> > told to use openMPI for this. >> > i'm not familiar w/ openMPI so i did an apt install of libopenmpi-dev. >> > >> > Do i have to loop through my job submission w/ mpirun >> > and run an sbatch outside of it? >> > Again, i'm still new to this, and w/ sge it was pretty straight fwd >> > where all i had to do was: >> > loop through files >> > qsub -N {name of job} script.sh {filename}. >> > >> > not sure how would i do so here. >> >>