Re: [R] Large datasets in R

From: Gabor Grothendieck <>
Date: Tue 18 Jul 2006 - 05:10:07 EST

You may or may not have problems. R keeps its data in memory so you will have to have sufficient memory to hold the data plus all derived data and code. Since R is free you can try it out. If your problems are
too large you can always get more memory or use S-Plus which can handle larger datasets and the code is similar to R so you can largely reuse your code.

On 7/17/06, Deepankar Basu <> wrote:
> Hi!
> I am a student of economics and currently do most of my statistical work
> using STATA. For various reasons (not least of which is an aversion for
> proprietary software), I am thinking of shifting to R. At the current
> juncture my concern is the following: would I be able to work on
> relatively large data-sets using R? For instance, I am currently working
> on a data-set which is about 350MB in size. Would be possible to work
> data-sets of such sizes using R?

> I have been trying to read up the posting on the R-archive on this
> topic; but I could not really understand all the discussion, nor could I
> reach the "end". So, I am not aware of the current state of consensus on
> the issue.
> It would help a lot if some current user could throw some light on this
> issue of large data-sets in R.
> Thanks in advance.
> Deepankar Basu
> ______________________________________________
> mailing list
> PLEASE do read the posting guide!
> mailing list PLEASE do read the posting guide! Received on Tue Jul 18 05:17:48 2006

Archive maintained by Robert King, hosted by the discipline of statistics at the University of Newcastle, Australia.
Archive generated by hypermail 2.1.8, at Tue 18 Jul 2006 - 06:16:44 EST.

Mailing list information is available at Please read the posting guide before posting to the list.