I have an R workflow that now needs to merge some large csv files into a single file. I find I am running out of memory so the merge has started crashing - these files are getting larger each month so it needs a fix. I could write some other script like Perl or whatever, but I would prefer to keep the whole workflow in R. It's on windows so some of the cmd script solutions would be horrible (and hard to check the formats as I go).
I've been looking around to see how to read files in chunks using R and your library seems perfect. So, I have a question more than an issue....
Do you think it would be possible to use it to merge files (in this case it is a known set of file names each time). I can't yet see a way to do it.
thanks very much for a useful library, we are now using it to filter extracts out of some very large source files and it's simplified quite a bit of our data cleaning/processing already.
I have an R workflow that now needs to merge some large csv files into a single file. I find I am running out of memory so the merge has started crashing - these files are getting larger each month so it needs a fix. I could write some other script like Perl or whatever, but I would prefer to keep the whole workflow in R. It's on windows so some of the cmd script solutions would be horrible (and hard to check the formats as I go).
I've been looking around to see how to read files in chunks using R and your library seems perfect. So, I have a question more than an issue....
Do you think it would be possible to use it to merge files (in this case it is a known set of file names each time). I can't yet see a way to do it.
thanks very much for a useful library, we are now using it to filter extracts out of some very large source files and it's simplified quite a bit of our data cleaning/processing already.