Message-ID: <484073.27472.qm@web32207.mail.mud.yahoo.com>
Date: 2007-12-18T07:02:03Z
From: Moshe Olshansky
Subject: read.table() and precision?
In-Reply-To: <6d32d7830712170934m4f9aa742je29da95b7eaab869@mail.gmail.com>
Dear List,
Following the below question I have a question of my
own:
Suppose that I have large matrices which are produced
sequentially and must be used sequentially in the
reverse order. I do not have enough memory to store
them and so I would like to write them to disk and
then read them. This raises two questions:
1) what is the fastest (and the most economic
space-wise) way to do this?
2) functions like write, write.table, etc. write the
data the way it is printed and this may result in a
loss of accuracy. Is there any way to prevent this,
except for setting the "digits" option to a higher
value or using format prior to writing the data? Is it
possible to write binary files (similar to Fortran)?
Any suggestion will be greatly appreciated.
--- Wojciech Gryc <wojciech at gmail.com> wrote:
> Hi,
>
> I'm currently working with data that has values as
> large as 99,000,000
> but is accurate to 6 decimal places. Unfortunately,
> when I load the
> data using read.table(), it rounds everything to the
> nearest integer.
> Is there any way for me to preserve the information
> or work with
> arbitrarily large floating point numbers?
>
> Thank you,
> Wojciech
>
> --
>
> Five Minutes to Midnight:
> Youth on human rights and current affairs
> http://www.fiveminutestomidnight.org/
>
> ______________________________________________
> R-help at r-project.org mailing list
> https://stat.ethz.ch/mailman/listinfo/r-help
> PLEASE do read the posting guide
> http://www.R-project.org/posting-guide.html
> and provide commented, minimal, self-contained,
> reproducible code.
>