I have pretty big data sizes, like matrices of .5 to 1.5GB so once i need to juggle several of them i am in need of disk cache. I am trying to use bigmemory package but getting problems that are hard to understand. I am getting seg faults and machine just hanging. I work by the way on Red Hat Linux, 64 bit R version 10.
Simplest problem is just saving matrices. When i do something like r<-matrix(rnorm(100), nr=10); librarybigmemory) for(i in 1:3) xx<-as.big.matrix(r, backingfile=paste("r",i, sep="", collapse=""), backingpath=MyDirName) it works just fine -- saves small matrices as three different matrices on disc. However, when i try it with real size, like with r<-matrix(normal(50000000), nr=1000) I am either getting seg fault on saving the third big matrix, or hang forever. Am i doing something obviously wrong, or is it an unstable package at the moment? Could anyone recommend something similar that is reliable in this case? -- View this message in context: http://r.789695.n4.nabble.com/bigmemory-package-woes-tp2062996p2062996.html Sent from the R help mailing list archive at Nabble.com. ______________________________________________ R-help@r-project.org mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code.