R-beta: Memory requirement in Win NT

Joe Mortzheim (jmortz@snake1.cr.usgs.gov)
Tue, 7 Apr 1998 07:07:42 -0400


From: Joe Mortzheim <jmortz@snake1.cr.usgs.gov>
To: "'r-help@stat.math.ethz.ch'" <r-help@stat.math.ethz.ch>
Subject: R-beta: Memory requirement in Win NT
Date: Tue, 7 Apr 1998 07:07:42 -0400

I generally use S+ but have been dismayed by it's slow operation and =
memory hogging.  I recently downloaded R and was hoping my memory =
problems were over.  To the contrary, R seems to run out of memory =
almost immediately; particularly during read.table().

I tried to change the arguments on startup of R, for example:

	R -n200000 -v20

However, I get an error:

	The instruction at "0x0008001f" referenced memory at "0x0008001f".  The =
memory could not be "read"

I get this error regardless of the values I use for -n or -v and I get =
the error if I try to use either the -n OR the -v arguments.

What gives?  It's enough to make me switch to SAS or (God forbid!) begin =
longing for a Microsoft product.

I have a Pentium 166 with 64 Megs of RAM and I'm running Windows NT.  =
Surely this is enough computational power if only I could tell R to use =
more resources than it starts with as default?  I would rather not go =
through the hassle of breaking down a read.table() command into a bunch =
of scan() commands or break up my dataset into multiple files.  I can =
load the same dataset with read.table() using S+ on this machine, it =
just takes forever.

I certainly appreciate any help.  Please respond to =
jmortz@snake1.cr.usgs.gov as well as the mailing list.

Thanks in advance for the advice.
-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-
r-help mailing list -- Read http://www.ci.tuwien.ac.at/~hornik/R/R-FAQ.html
Send "info", "help", or "[un]subscribe"
(in the "body", not the subject !)  To: r-help-request@stat.math.ethz.ch
_._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._