Hi Matt, you're correct: length(ffObject) must be smaller than 2^31-1... at least until R has a 64bit integer type, it seems...
in the meantime, use the bigmemory package. ;-) b On Fri, Feb 5, 2010 at 4:18 PM, Matthew Keller <mckellerc...@gmail.com> wrote: > Hello all, > > I hate to add to the daily queries regarding R's handling of large > datsets ;), but... > > I read in an online powerpoint about the ff package something about > the "length of an ff object" needing to be smaller than > .Machine$integer.max. Does anyone know if this means that the # of > elements in an ff object must be < .Machine$integer.max [i.e., that ff > provides no help with respect to the number of elements in a given > object]? I've got a matrix that has 19e9 elements and - even though I > can fit it into my ram (using "raw" storage.mode) - R won't let me > store it because 19e9 is >> .Machine$integer.max = 2^31. > > Anyone else have suggestions on how to deal with such massive datasets > like the ones I'm using? I'm exploring ncdf as we speak. Best, > > Matt > > > > -- > Matthew C Keller > Asst. Professor of Psychology > University of Colorado at Boulder > www.matthewckeller.com > > ______________________________________________ > R-help@r-project.org mailing list > https://stat.ethz.ch/mailman/listinfo/r-help > PLEASE do read the posting guide http://www.R-project.org/posting-guide.html > and provide commented, minimal, self-contained, reproducible code. > ______________________________________________ R-help@r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code.