Skip to content
Prev 4133 / 15274 Next

Preprocessing RData file (data.table and ff, bigmemory)

You could also have a look at filehash, with which I've been playing around for a while. 

Furthermore, I recently found a package still in Beta version, colbycol, written by Carlos J. Gil Bellosta (http://www.datanalytics.com) which seems to do what's on the tin: reading and managing large datasets well beyond the in-process memory limits...

Jos?

-----Original Message-----
From: r-sig-finance-bounces at stat.math.ethz.ch [mailto:r-sig-finance-bounces at stat.math.ethz.ch] On Behalf Of Steve Jaffe
Sent: 22 May 2009 15:19
To: r-sig-finance at stat.math.ethz.ch
Subject: Re: [R-SIG-Finance] [R-sig-finance] Preprocessing RData file (data.table and ff, bigmemory)


I'm new to R and interested in working with large amounts of data
(timeseries, but regularly spaced.) Can you point me to a good reference for
using data.table with bigmemory or ff? 

(I'm a bit puzzled about what exactly these packages provide. As I
understand it, on 32-bit platforms files are subject to the same 2GB limit
as in-process memory, so I assume that dealing with a larger dataset still
requires breaking it up into multiple files...)

Thanks for your help.



I failed to point out that data.table can make use of both (?) those
packages. [ff, bigmemory]

It isn't a time-series library per se, but it make one very cool
in-memory database.  Similar in spirit to some of the not-so-free ones
out there...

Jeff