Replies: 1 comment
-
I think the larger the chunk size the better the performance
Try removing the above or set it to much higher number e.g. Are the files share-able? If yes I can take a look. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I am fairly new at handling medium sized data so I could very well be doing something basic wrong, but I am not seeing what my issue could be.
I have a 9 GB csv file, and am running on an 11th Gen Intel(R) Core(TM) i7-1185G7 @ 3.00 Ghz with 4 cores, 8 logical processors, and 15.7 GB of Memory. I am using R 3.6.1 (can't update due to employer). 120M rows and 19 columns.
When I run the fallowing code the stage1splitter runs for hours with no results. and the cpu usage for the R for Windows front-end workers is 0% most of the time.
Code
Output
Is this an issue with using 3.6.1? When I load disk.frame I get
Warning message: package ‘disk.frame’ was built under R version 3.6.3
Beta Was this translation helpful? Give feedback.
All reactions