Re: [R] memory

From: Prof Brian Ripley <>
Date: Tue 30 Aug 2005 - 17:12:10 EST

On Mon, 29 Aug 2005, Ferran Carrascosa wrote:

> Hi,
> I have a matrix with 700.000 x 10.000 cells with floating point data.
> I would like to work with the entire table but I have a lot of memory
> problems. I have read the ?memory
> I work with Win 2000 with R2.1.0
> The only solution that I have applied is:
>> memory.limit(size=2048)
> But now my problems are:
> - I need to work with more than 2 Gb. How I can exceed this limit?

Re-read the rw-FAQ, or (preferably) get a more capable OS on a 64-bit CPU.

> - When apply some algorithms, the maximum cells in one object 2*10^9
> (aprox.) is reached.

You will never get that many cells (that is the address space in bytes, and they are several bytes each). Please do as the posting guide asks and report accurately what happened.

> Please could you send me some advises/strategies about the work with
> large amount of data in R?
> R have a way to work with less memory needs?

Your matrix has 7e09 cells (assuming you are using . as a thousands separator) and needs 5.6e10 bytes to store. Your OS has a memory address limit of 3.2e09 bytes. Don't blame R for being limited by your OS.

Brian D. Ripley,        
Professor of Applied Statistics,
University of Oxford,             Tel:  +44 1865 272861 (self)
1 South Parks Road,                     +44 1865 272866 (PA)
Oxford OX1 3TG, UK                Fax:  +44 1865 272595

______________________________________________ mailing list
PLEASE do read the posting guide!
Received on Tue Aug 30 17:16:50 2005

This archive was generated by hypermail 2.1.8 : Fri 03 Mar 2006 - 03:40:00 EST