How to load a big txt file
Easy solution will be split your big txt files by text editor. e.g. 5000 rows each. and then combine the dataframes together into one.
On 6/7/07, ssls sddd <ssls.sddd at gmail.com> wrote:
Dear list,
I need to read a big txt file (around 130Mb; 23800 rows and 49 columns)
for downstream clustering analysis.
I first used "Tumor <- read.table("Tumor.txt",header = TRUE,sep = "\t")"
but it took a long time and failed. However, it had no problem if I just put
data of 3 columns.
Is there any way which can load this big file?
Thanks for any suggestions!
Sincerely,
Alex
[[alternative HTML version deleted]]
______________________________________________ R-help at stat.math.ethz.ch mailing list https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide http://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code.
"The scientists of today think deeply instead of clearly. One must be sane to think clearly, but one can think deeply and be quite insane." Nikola Tesla http://www.macgrass.com