On Tue, 17 Apr 2001 cdhershberger@dow.com wrote:
Full_Name: Doug Hershberger Version: 1.2.2 OS: Red Hat-Linux 7.0 Submission from: (NULL) (216.99.65.36) In the R FAQ I find the following entry explaining that R no longer has problems with memory.
No longer has the same problems.... There is also a section on `What is a bug?' that you should read again.
http://cran.r-project.org/doc/FAQ/R-FAQ.html#Why%20does%20R%20run%20out%20of%20of%20memory%3f However in my installation: R Version 1.2.2 (2001-02-26) Installed from the red hat RPM on your site on a Red Hat 7.0 i686 I get the following error when working with large data sets:
> source("/usr/local/genex/rcluster/lib/rcluster/r/hcluster.r");
> breadth.program("uploaded_data.txt", "average", 10)
Read 2 items
Read 8574 items
Error: cannot allocate vector of size 71773 Kb
Execution halted
Is there a way to fix this? This is running as part of the GeneX
Install more memory, or write better R code, or use better statistical methods.
installation so I would have to dig through to figure out how to give R more memory. Besides which I don't think that will work.
We would need to know a lot more about the code to see the problem, but as a quick guess, clustering 8574 items with the standard algorithms takes a very large amount of memory (on any system), so it would be the third alternative.
Brian D. Ripley, ripley@stats.ox.ac.uk Professor of Applied Statistics, http://www.stats.ox.ac.uk/~ripley/ University of Oxford, Tel: +44 1865 272861 (self) 1 South Parks Road, +44 1865 272860 (secr) Oxford OX1 3TG, UK Fax: +44 1865 272595 -.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.-.- r-devel mailing list -- Read http://www.ci.tuwien.ac.at/~hornik/R/R-FAQ.html Send "info", "help", or "[un]subscribe" (in the "body", not the subject !) To: r-devel-request@stat.math.ethz.ch _._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._._