gt; > > -Dj
> > >
> > >
> > > ___
> > > users mailing list
> > > users@gridengine.org
> > > https://gridengine.org/mailman/listinfo/users
> >
> > --
> > > Andreas Haupt|
And do you define host groups in the PE?
On Tue, Mar 12, 2019 at 9:53 AM David Trimboli wrote:
>
> On 3/12/2019 12:05 PM, Ian Kaufman wrote:
> > Are mynode{17-24} in a queue that is configured to use your "threads" PE?
>
>
> Yes. If you disable the limit, the subm
ail:
>
> qsub -l vf=1G -l
> h="mynode17|mynode18|mynode19|mynode20|mynode21|mynode22|mynode23|mynode24"
> -pe threads 1 anyscript.sh
>
> It'll work if you remove "-pe threads 1".
> ___
> users mailing l
IO issues? NFS server providing data and possibly jobs running over NDS
shares as opposed to running on local disk?
On Sat, Jan 26, 2019, 11:23 AM Joseph Farran Hi Daniel.
>
> Yes I do have large job-arrays around 7k tasks BUT I have had larger job
> arrays of 500k without seeing this kind of slo
helped, you may need to
> > unload/fix the apparmor profiles already loaded in the kernel.
> >
> > I haven't worked with apparmor, so you may need to search the internet
> > for solutions.
>
>
> I'll take a look. My brief perusal suggests that it's a nontrivial task
> to do so. I'll see what I find.
>
> Thanks!
>
> ___
> users mailing list
> users@gridengine.org
> https://gridengine.org/mailman/listinfo/users
>
--
Ian Kaufman
Research Systems Administrator
UC San Diego, Jacobs School of Engineering ikaufman AT ucsd DOT edu
___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users
and it's not really fair for other user that use qsub.
>
> Then the simpliest way avoiding this is to forbiden the connection
>
> I'm not specialist of PAM and authentication have you got a link ?
>
>
>
>
> ___
>
edu)
> -- Genome Sciences Department, System Administrator
> -- Foege Building S046, (206)-685-7354
> -- University of Washington School of Medicine
> _______
> users mailing list
> users@gridengine.org
> https://grideng
t; weight_tickets_functional 1
> weight_tickets_share 0
>
> Perhaps these settings might be causing our issue? Seems unlikely though,
> as we're not taking project or department into account in our scheduling.
>
> Thanks,
>
> DR
>
>
> On 2017-10-
bit of googling and man page
> reading on the relevant topics and settings, but wasn't able to find a good
> explanation for the behavior we're seeing. Any help greatly appreciated!
>
> Thanks,
>
> DR
> __
ault)
>
>
> Thanks,
> Subashini.K
>
> ___
> users mailing list
> users@gridengine.org
> https://gridengine.org/mailman/listinfo/users
>
>
--
Ian Kaufman
Research Systems Administrator
UC San Diego, Ja
ease
> notify the sender immediately and delete this email from your computer.
>
> ___
> users mailing list
> users@gridengine.org
> https://gridengine.org/mailman/listinfo/users
>
>
--
Ian Kaufman
Research Systems Administrator
UC San Diego, Jacobs School of Engineering ikaufman AT
or message ***
>
>
> /opt/gridengine/default/spool/compute-0-4/job_scripts/5020911: line 12:
> 9088 Segmentation fault (core dumped) mpirun -np 4 gmx mdrun -ntmpi 1
> -ntomp 8 -v -deffnm eql2
>
>
>
> What is the reason behind this?
>
>
> Thanks,
> Subashi
s error?
>
>
> Please help.
>
> Thanks,
> Subashini.K
>
> ___
> users mailing list
> users@gridengine.org
> https://gridengine.org/mailman/listinfo/users
>
>
--
Ian Kaufman
Resear
Executable: /usr/local/gromacs/bin/gmx
>>> Library dir: /usr/local/gromacs/share/gromacs/top
>>>
>>>
>>> Can anyone help me?
>>>
>>> Thanks,
>>> Subashini.K
>>>
>>>
>>> --
>&
___
> users mailing list
> users@gridengine.org
> https://gridengine.org/mailman/listinfo/users
>
--
Ian Kaufman
Research Systems Administrator
UC San Diego, Jacobs School of Engineering ikaufman AT ucsd DOT edu
_
oups who are in child domains like
> NAFTA.COMPANY.COM and EAME.COMPANY.COM etc.
>
> -dag
>
>
>
> Ian Kaufman wrote:
>
>> If you support multiple domains, are you able to guarantee unique short
>> names? It seems to me that could be a problem. If it is a case of multiple
&
;>
>
> ___
> users mailing list
> users@gridengine.org
> https://gridengine.org/mailman/listinfo/users
>
--
Ian Kaufman
Research Systems Administrator
UC San Diego, Jacobs School of Engineering ikaufman AT ucsd DOT edu
___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users
for the use of the named
> recipient(s) above. Any unauthorized use or disclosure of this email is
> strictly prohibited. If you are not the intended recipient(s), please
> notify the sender immediately and delete this email from your computer.
>
> ___
> users mailing list
> users@grideng
:50
>>> 1
>>> 1144482 0.55500 sas64 username Eqw 09/27/2016 23:30:40
>>> 1
>>> 1144484 0.55500 sas64 username Eqw 09/27/2016 23:31:30
>>>
ially if anybody is trying to do some kind of visualization
> on them, but I've never had complaints about them being under-powered yet.
>
> Any thoughts you might have are appreciated.
>
> Thanks
> Biggles
>
> _______
> users
gt;> ___
>>> users mailing list
>>> users@gridengine.org
>>> https://gridengine.org/mailman/listinfo/users
>>>
>>>
>> --
>> Alex Chekholko ch...@stanford.edu
>>
>> _
h memory allocated and then use qacct
> to look at the actual max usage so they know what they should ask for next
> time. We had some teething troubles with this for a few weeks after it was
> introduced, but it's all been working smoothly for a long time now.
>
>
--
Ian Kaufman
> management behind the scenes.
>
> Chris
>
>
>
> Pat Haley wrote:
> >
> > It looks similar but one big difference is when I run "qconf -sh" I
> > see all my compute nodes listed along with my frontend. However
> > "qconf -sql"
of the sender and do not necessarily represent the views of the
>>> Babraham Institute. Full conditions at:
>>> www.babraham.ac.uk<http://www.babraham.ac.uk/terms>
>>>
>>> ___
>&g
t basis?) so jobs that don't explicitly request the resource will use
> something by default, but we've been unable to figure out how to set this
> up.
>
> We are using 6.1u3.
>
> thanks
>
> ___________
> users mailing list
> 0.0
> Host Resource(s): hl:mem_free=88.885G
>
> And everything was working until a week ago.
>
> Ilya.
>
> Original Message
> Subject: Re: [gridengine users] Cannot request resource if it is a load
> value of memory type: SGE reports it as unk
-
>>>>>>> -
>>>>>>> gpu038 lx24-amd64 16 2.11 126.1G 15.7G
>>>>>>> 4.0G 0.0
>>>>>>> Host Resource(s): hl:mem_free=110.416G
>>>>>>>
>>>>>>>
g list
> users@gridengine.org
> https://gridengine.org/mailman/listinfo/users
>
>
>
> --
> Dr. James Abbott
> Lead Bioinformatician
> Bioinformatics Support Service
> Imperial College, London
>
>
> ___
> users maili
a user to use more than 3 entire nodes.
>
> Maybe there is some sorts of resource equivalency between slot and memory can
> achieve that?
>
> Thanks
> D
>
> Sent from my iPad
>
>> On 1 Jul 2014, at 5:57 am, Ian Kaufman wrote:
>>
>> I don't get th
ive or take. So it makes it really impracticle actually.
> However so far I've just set the default h_vmem and s_vmem values high
> enough to accomadate jvm apps, and increased the per-host consumable
> appropriately. We don't get fine-grained memory control, but it definitely
>
If not, you may want to limit the number of jobs to the
number of cores/slots you have, otherwise you will run into a backlog
of jobs, and if an array job, this can even be more serious.
Ian
--
Ian Kaufman
Research Systems Administrator
UC San Diego, Jacobs School of Engineering ikaufman AT
L
> parser, or is there a more convenient way to get the host?
>
> Tim
>
> ___
> users mailing list
> users@gridengine.org
> https://gridengine.org/mailman/listinfo/users
--
Ian Kaufman
Rese
; Yes and no. This is clearly a client-side configuration error that is
> impacting GE, but if GE is the only application/service that can't get user
> information correctly, maybe there is a problem with GE. Ultimately, I was
> just asking for help in figuring out what is misconfigu
gives the signal.
>> > is there something I'm missing as to how maxvmem is calculated?
>>
>> SGE will not look at one process only, but at the complete process tree.
>> Is PhythonRender the one and only child of `sge_shepherd`?
>>
>> -- Reuti
>
>
>
> ___
> users mailing list
> users@gridengine.org
> https://gridengine.org/mailman/listinfo/users
>
--
Ian Kaufman
Research Systems Administrator
UC San Diego, Jacobs School of Engineering ikaufman AT ucsd DOT edu
___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users
Nyhuis
> IT Manager, Stam Lab
> 2211 Elliott Avenue
> 6th Floor, 6S139
> Seattle, WA 98121
> O: (206)-267-1097 ext 220
> F: (206)-441-3033
>
>
>
> ___
> users mailing list
> users@gridengine.org
> https://gridengine.o
And here is some more info:
http://serverfault.com/questions/322073/howto-set-up-sge-for-cuda-devices
On Mon, Apr 14, 2014 at 1:39 PM, Ian Kaufman wrote:
> If everything is configured correctly, GridEngine will be aware that
> the GPU in node1 is in use, and schedule around it, ensurin
If everything is configured correctly, GridEngine will be aware that
the GPU in node1 is in use, and schedule around it, ensuring that the
8 GPU job will get unused GPUs.
Ian
On Mon, Apr 14, 2014 at 1:38 PM, Ian Kaufman wrote:
> Look at the info presented here:
>
> http://stackove
s are still free for jobs.
>
> I submit one job with PE=8. SGE allocate all the 3 nodes to me with 8
> GPU slots. The problem is now: how my job knows what GPUs it can get
> on node1?
>
> Best
>
>
>
>
> On Mon, Apr 14, 2014 at 4:13 PM, Ian Kaufman wrote:
>>
t;
>> Like the option to request sockets instead of cores which I posted in the
>> last couple of days, you can use a similar approach to get the number of the
>> granted GPU out of the queue name.
>>
>> -- Reuti
>>
>>
>>> On Mon, Apr 14, 2014 at
olution, since some jobs may still happen to collide to each
> other on the same GPU on a multiple GPU node. If GE can have the
> memory to record the GPUs allocated to a job, then this can be
> perfect.
>
>
> On Mon, Apr 14, 2014 at 1:46 PM, Ian Kaufman wrote:
>> I belie
and manage GPU(and Phi). Not need to be complicated and
> powerful, just do basic work.
>
> Thanks,
> ___
> users mailing list
> users@gridengine.org
> https://gridengine.org/mailman/listinfo/users
--
Ian Kaufman
Research Sys
2.118G
>
> and I requesting the wrong resource? how gridengine measured it's 2.118G is
> unclear to me.
>
>
> ___________
> users mailing list
> users@gridengine.org
> https://gridengine.org/mailman/listinfo/users
>
--
Ian Kaufman
Research Systems Administrator
U
ovation Campus - 01235 77 8442
>
> --
> This e-mail and any attachments may contain confidential, copyright and or
> privileged material, and are for the use of the intended addressee only. If
> you are not the intended addressee or an authorised recipient of the
> addressee please notify us of receipt by returning the e-mail and do not
> use, copy, retain, distribute or disclose the information in or attached to
> the e-mail.
> Any opinions expressed within this e-mail are those of the individual and
> not necessarily of Diamond Light Source Ltd. Diamond Light Source Ltd.
> cannot guarantee that this e-mail or any attachments are free from viruses
> and we cannot accept liability for any damage which you may sustain as a
> result of software viruses which may be transmitted in or with the message.
> Diamond Light Source Limited (company no. 4375679). Registered in England
> and Wales with its registered office at Diamond House, Harwell Science and
> Innovation Campus, Didcot, Oxfordshire, OX11 0DE, United Kingdom
>
>
>
>
> ___
> users mailing list
> users@gridengine.org
> https://gridengine.org/mailman/listinfo/users
--
Ian Kaufman
Research Systems Administrator
UC San Diego, Jacobs School of Engineering ikaufman AT ucsd DOT edu
___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users
; Regards
>>
>> Lionel
>> ___
>> users mailing list
>> users@gridengine.org
>> https://gridengine.org/mailman/listinfo/users
>
> ___
> users mailing list
> users@gridengine.org
>
org/mailman/listinfo/users
>>
>
>
> --
> Richard Ems mail: richard@cape-horn-eng.com
>
> Cape Horn Engineering S.L.
> C/ Dr. J.J. Dómine 1, 5º piso
> 46011 Valencia
> Tel : +34 96 3242923 / Fax 924
> http://www.cape-horn-eng.com
> __
n Technology
> http://www.oit.uci.edu/rcs/
>
> University of California, Irvine
> www.ics.uci.edu/~aebrenne/
> aebre...@uci.edu
> _______
> users mailing list
> users@gridengine.org
> https://gridengine.org/mailman/listinfo/users
>
--
Ian Kaufman
Res
in this case it is neither. The host
> only has this one queue instance, and the job only requests this one
> complex.
>
> Suggestions?
>
> Regards,
> --
> Alex Chekholko ch...@stanford.edu
> __**_
> users mailing list
> users@gridengi
01 Spruce St.
>
> Room 214
>
> Philadelphia PA 19104
>
> Tel: 215-495-6914
>
> Sending me a large file? Use my secure dropbox:
>
> https://cscb-filetransfer.wistar.upenn.edu/dropbox/btay...@wistar.org
>
> ** **
>
> *From:* Ian Kaufman [ma
**
>
> Brett Taylor
>
> Systems Administrator
>
> Center for Systems and Computational Biology
>
>
> The Wistar Institute
>
> 3601 Spruce St.
>
> Room 214
>
> Philadelphia PA 19104
>
> Tel: 215-495-69
s comments - there would be no "$" preceding the "\r" since the very
next character after the newline gets interpreted is the "\r".
Ian
--
Ian Kaufman
Research Systems Administrator
UC San Diego, Jacobs School of Engineering ikaufman AT ucsd DOT edu
___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users
h
>>>
>>> #$-N SAFKTAT1
>>> #$-t 6970-10230:1
>>>
>>>
>>>
>>> /var/software/packages/Mathematica/7.0/Executables/math -run
>>> "teller=$SGE_TASK_ID;<< ModelFotokatalyseTAT.m"
>>>
>>>
>>
>
> ___
> users mailing list
> users@gridengine.org
> https://gridengine.org/mailman/listinfo/users
>
>
--
Ian Kaufman
Research Systems Administrator
UC San Diego, Jacobs School of Engineering ikaufman AT ucsd DOT edu
___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users
RAM consumed at any one point in time up until the
moment you ran the qstat command.
Ian
--
Ian Kaufman
Research Systems Administrator
UC San Diego, Jacobs School of Engineering ikaufman AT ucsd DOT edu
___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users
s of production and just recently over 100K jobs in two
months).
Ian
--
Ian Kaufman
Research Systems Administrator
UC San Diego, Jacobs School of Engineering ikaufman AT ucsd DOT edu
___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users
ithout mucking with any flags. Additionally,
the crew I work with use a workflow/job management tool that isn't
thread aware, so every child Java app that gets launched gets its own
JVM. That is high on the list of things to fix!
Ian
--
Ian Kaufman
Research Systems Administrator
UC San D
ws me to ensure that on a 12 core box
and 48GB of RAM, I can run nearly 12 jobs keeping the desired 1 per
core.
Ian
--
Ian Kaufman
Research Systems Administrator
UC San Diego, Jacobs School of Engineering ikaufman AT ucsd DOT edu
___
users mailin
ault, Grid Engine uses one slot per
core, and one job fits in each slot.
Ian
--
Ian Kaufman
Research Systems Administrator
UC San Diego, Jacobs School of Engineering ikaufman AT ucsd DOT edu
___
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users
rectly limits user alpha to 320 jobs, user beta to 128 jobs,
and all others to 32 jobs. In your rqs, you are limiting users to 16
jobs, not 2.
Ian
--
Ian Kaufman
Research Systems Administrator
UC San Diego, Jacobs School of Engineering ikaufman AT ucsd DOT edu
___
57 matches
Mail list logo