Hi again and thank you all for the answers

i need to add that im a relatively R neewb so i apologize in advance

i started R with the --vanilla option and ran gc()

this is the output i get:

> gc()
         used (Mb) gc trigger (Mb) max used (Mb)
Ncells 182236  9.8     407500 21.8   350000 18.7
Vcells 277896  2.2     786432  6.0   785897  6.0

also this is the memory.profile()

> memory.profile()
       NULL      symbol    pairlist     closure environment     promise
          1        5611       86695        2277         314        4175
   language     special     builtin        char     logical     integer
      21636          44         637        6361        4574       11089
     double     complex   character         ...         any        list
        782           1       20934           0           0        8023
 expression    bytecode externalptr     weakref         raw          S4
          1        6271        1272         364         365         831
>

im running on linux (arch linux) and 'free' shows this:


zeltak@zuni ~ ↳ free -h
             total       used       free     shared    buffers     cached
Mem:          251G        99G       152G        66G       249M        84G
-/+ buffers/cache:        14G       237G
Swap:           0B         0B         0B

im not running any parrallel stuff at all

milan: how does one know if the  memory is fragmented?

thank you all again i really appreciate the help

best

Z



On Thu, Jan 2, 2014 at 10:35 PM, Ben Bolker <bbol...@gmail.com> wrote:

> Xebar Saram <zeltakc <at> gmail.com> writes:
>
> >
> > Hi All,
> >
> > I have a terrible issue i cant seem to debug which is halting my work
> > completely. I have R 3.02 installed on a linux machine (arch
> linux-latest)
> > which I built specifically for running high memory use models. the system
> > is a 16 core, 256 GB RAM machine. it worked well at the start but in the
> > recent days i keep getting errors and crashes regarding memory use, such
> as
> > "cannot create vector size of XXX, not enough memory" etc
> >
> > when looking at top (linux system monitor) i see i barley scrape the 60
> GB
> > of ram (out of 256GB)
> >
> > i really don't know how to debug this and my whole work is halted due to
> > this so any help would be greatly appreciated
>
>   I'm very sympathetic, but it will be almost impossible to debug
> this sort of a problem remotely, without a reproducible example.
> The only guess that I can make, if you *really* are running *exactly*
> the same code as you previously ran successfully, is that you might
> have some very large objects hidden away in a saved workspace in a
> .RData file that's being loaded automatically ...
>
>   I would check whether gc(), memory.profile(), etc. give sensible results
> in a clean R session (R --vanilla).
>
>   Ben Bolker
>
> ______________________________________________
> R-help@r-project.org mailing list
> https://stat.ethz.ch/mailman/listinfo/r-help
> PLEASE do read the posting guide
> http://www.R-project.org/posting-guide.html
> and provide commented, minimal, self-contained, reproducible code.
>

        [[alternative HTML version deleted]]

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to