Hi Reuti I solved this problem, I tried reinstall without GUI and It's success.
Thanks for your help! Regards! On 7/29/14, Reuti <re...@staff.uni-marburg.de> wrote: > Hi, > > Am 29.07.2014 um 06:07 schrieb Derrick Lin: > >> This is qhost of one of our compute nodes: >> >> pwbcad@gamma01:~$ qhost -F -h omega-0-9 >> HOSTNAME ARCH NCPU LOAD MEMTOT MEMUSE SWAPTO >> SWAPUS >> ------------------------------------------------------------------------------- >> global - - - - - - >> - >> omega-0-9 lx26-amd64 64 12.34 504.9G 273.6G 256.0G >> 14.6G >> hl:arch=lx26-amd64 >> hl:num_proc=64.000000 >> hl:mem_total=504.890G >> hl:swap_total=256.000G >> hl:virtual_total=760.890G >> hl:load_avg=12.340000 >> hl:load_short=9.720000 >> hl:load_medium=12.340000 >> hl:load_long=18.900000 >> hl:mem_free=231.308G >> hl:swap_free=241.356G >> hl:virtual_free=472.663G >> hl:mem_used=273.582G >> hl:swap_used=14.644G >> hl:virtual_used=288.226G >> hl:cpu=15.400000 >> >> hl:m_topology=SCTTCTTCTTCTTCTTCTTCTTCTTSCTTCTTCTTCTTCTTCTTCTTCTTSCTTCTTCTTCTTCTTCTTCTTCTTSCTTCTTCTTCTTCTTCTTCTTCTT >> >> hl:m_topology_inuse=SCTTCTTCTTCTTCTTCTTCTTCTTSCTTCTTCTTCTTCTTCTTCTTCTTSCTTCTTCTTCTTCTTCTTCTTCTTSCTTCTTCTTCTTCTTCTTCTTCTT >> hl:m_socket=4.000000 >> hl:m_core=32.000000 >> hl:np_load_avg=0.192812 >> hl:np_load_short=0.151875 >> hl:np_load_medium=0.192812 >> hl:np_load_long=0.295312 >> hc:mem_requested=502.890G > > So, here is no h_vmem on an exechost level. > > >> We do not set h_vmem in queue instance level, that's intended because we >> just need h_vmem in per user quota like: > > Typo and you mean exechost level? > > >> { >> name default_per_user >> enabled true >> description "Each user entitles to resources equivalent to two >> nodes" >> limit users {*} queues {all.q} to slots=16,h_vmem=16G >> } > > RQS limits are not enforced. The user has to specify it by hand then with > the -l option to `qsub`. > > Is "h_vmem" then in "complex_values" in the queue definition with an initial > value per queue instance? > > -- Reuti > > >> At the queue instance level, we use mem_requested as "per host quota" >> instead. It's a custom complex attr we setup for our specific >> applications. >> >> Cheers, >> D >> >> >> On Tue, Jul 29, 2014 at 1:02 AM, Reuti <re...@staff.uni-marburg.de> >> wrote: >> Hi, >> >> Am 04.07.2014 um 06:04 schrieb Derrick Lin: >> >> > Interestingly, I have a small test cluster that basically have the same >> > SGE setup does *not* have such problem. h_vmem in complex is exactly the >> > same. The test queue instance looks almost the same (except the CPU >> > layout etc) >> > >> > qstat -F -q all.q@eva00 >> > queuename qtype resv/used/tot. load_avg arch >> > states >> > --------------------------------------------------------------------------------- >> > all.q@eva00.local BP 0/0/8 0.00 lx26-amd64 >> > ... >> > hc:mem_requested=7.814G >> > qf:qname=all.q >> > qf:hostname=eva00.local >> > qc:slots=8 >> > qf:tmpdir=/tmp >> > qf:seq_no=0 >> > qf:rerun=0.000000 >> > qf:calendar=NONE >> > qf:s_rt=infinity >> > qf:h_rt=infinity >> > qf:s_cpu=infinity >> > qf:h_cpu=infinity >> > qf:s_fsize=infinity >> > qf:h_fsize=infinity >> > qf:s_data=infinity >> > qf:h_data=infinity >> > qf:s_stack=infinity >> > qf:h_stack=infinity >> > qf:s_core=infinity >> > qf:h_core=infinity >> > qf:s_rss=infinity >> > qf:h_rss=infinity >> > qf:s_vmem=infinity >> > qf:h_vmem=infinity >> > qf:min_cpu_interval=00:05:00 >> > >> > Both clusters don't have h_vmem defined in exechost level. >> >> What is the output of: >> >> `qhost -F` >> >> Below you write that it's also defined on a queue instance level, hence in >> both places (as "complex_values")? >> >> -- Reuti >> >> >> > Derrick >> > >> > >> > On Fri, Jul 4, 2014 at 1:58 PM, Derrick Lin <klin...@gmail.com> wrote: >> > Hi all, >> > >> > We start using h_vmem to control jobs by their memory usage. However >> > jobs couldn't start when there is -l h_vmem. The reason is >> > >> > (-l h_vmem=1G) cannot run in queue "intel.q@delta-5-1.local" because job >> > requests unknown resource (h_vmem) >> > >> > However, h_vmem is definitely on the queue instance: >> > >> > queuename qtype resv/used/tot. load_avg arch >> > states >> > --------------------------------------------------------------------------------- >> > intel.q@delta-5-1.local BIP 0/0/64 6.27 lx26-amd64 >> > .... >> > hl:np_load_long=0.091563 >> > hc:mem_requested=504.903G >> > qf:qname=intel.q >> > qf:hostname=delta-5-1.local >> > qc:slots=64 >> > qf:tmpdir=/tmp >> > qf:seq_no=0 >> > qf:rerun=0.000000 >> > qf:calendar=NONE >> > qf:s_rt=infinity >> > qf:h_rt=infinity >> > qf:s_cpu=infinity >> > qf:h_cpu=infinity >> > qf:s_fsize=infinity >> > qf:h_fsize=infinity >> > qf:s_data=infinity >> > qf:h_data=infinity >> > qf:s_stack=infinity >> > qf:h_stack=infinity >> > qf:s_core=infinity >> > qf:h_core=infinity >> > qf:s_rss=infinity >> > qf:h_rss=infinity >> > qf:s_vmem=infinity >> > qf:h_vmem=infinity >> > qf:min_cpu_interval=00:05:00 >> > >> > I tried to specify other attr such as h_rt, jobs started and finished >> > successfully. >> > >> > >> > >> > >> > qconf -sc >> > >> > >> > >> > #name shortcut type relop requestable consumable >> > default urgency >> > >> > >> > >> > #---------------------------------------------------------------------------------------- >> > >> > >> > >> > h_vmem h_vmem MEMORY <= YES YES >> > 0 0 >> > >> > >> > >> > # >> > >> > Can anyone shed light on this? >> > >> > Cheers, >> > Derrick >> > >> > _______________________________________________ >> > users mailing list >> > users@gridengine.org >> > https://gridengine.org/mailman/listinfo/users >> >> > > > _______________________________________________ > users mailing list > users@gridengine.org > https://gridengine.org/mailman/listinfo/users > -- Nguyen The Man Verification Team Phone Number: 0935.678.703 Email: man.nguyen...@icdrec.edu.vn _______________________________________________ users mailing list users@gridengine.org https://gridengine.org/mailman/listinfo/users