Re: [R] How to read HUGE data sets?

From: Gabor Grothendieck <ggrothendieck_at_gmail.com>
Date: Fri, 29 Feb 2008 08:39:51 -0500

On Fri, Feb 29, 2008 at 8:27 AM, Liviu Andronic <landronimirc_at_gmail.com> wrote:
> On 2/28/08, Gabor Grothendieck <ggrothendieck@gmail.com> wrote:
> > The sqldf package can read a subset of rows and columns (actually any
> > sql operation)
> > from a file larger than R can otherwise handle. It will automatically
> > set up a temporary
> > SQLite database for you, load the file into the database without going
> > through R and
> > extract just the data you want into R and then automatically delete
> > the database. All this
> > can be done in 2 lines of code.
>
> Is it realistic to use this approach for datasets as big as 30-40 GB?

The SQLite site says SQLite is appropriate up to a few dozen gigabytes. http://www.sqlite.org/whentouse.html

The only way to really know is to try it with your data. Since it does not involve much code it shouldn't take long to prepare a test.



R-help_at_r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code. Received on Fri 29 Feb 2008 - 13:59:37 GMT

Archive maintained by Robert King, hosted by the discipline of statistics at the University of Newcastle, Australia.
Archive generated by hypermail 2.2.0, at Fri 29 Feb 2008 - 14:00:18 GMT.

Mailing list information is available at https://stat.ethz.ch/mailman/listinfo/r-help. Please read the posting guide before posting to the list.

list of date sections of archive