Dear Berend,

thanks to everyone who responded. I should point that the data matrix is not 
sparse. Yes, the irlba package seemed of interest, however annoyingly, they 
don't specify in the package how large is "large"!!!, and when you read the 
paper on which it is based, it deals with rather smaller data matrices than the 
ones I have, so it is unclear how it would perform and scale.

So, no one has direct experience running irlba on a data matrix as large as 
500,000 x 1,000 or larger?

kind regards
Andy





________________________________
 From: Berend Hasselman <b...@xs4all.nl>

Cc: "r-help@R-project.org" <r-help@r-project.org> 
Sent: Monday, 8 April 2013, 20:31
Subject: Re: [R] SVD on very large data matrix




> 
> 
> Dear All,
> 
> I need to perform a SVD on a very large data matrix, of dimension ~ 500,000 x 
> 1,000 , and I am looking
> for an efficient algorithm that can perform an approximate (partial) SVD to 
> extract on the order of the top 50
> right and left singular vectors.
> 
> Would be very grateful for any advice on what R-packages are available to 
> perform such a task, what the RAM requirement is, and indeed what would be 
> the state-of-the-art in terms of numerical algorithms and programming
> language to use to accomplish this task.


Info found with package sos and findFn("svd") and scrolling through the list 
for something relevant.

Have a look at package irlba.
It can work with dense matrices and sparse matrices as provided by package 
Matrix, according to the documentation.

Berend
        [[alternative HTML version deleted]]

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to