-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Large datasets #1
Comments
Hi @tafia, first of all, congratulations on filing the first issue in the Up till now, the core C++ code of As you noticed, I have yet to write documentation on the
The future expansion features will be developed in the coming period using the IO operations using the Having a |
🥇
You sure have lot of work to do! I certainly don't want to bother you too much. I'll split my input file for the moment in as many chunks as necessary. For the moment, I am mainly interested in creating fst files (Write in-memory datasets and saving it to the disk). There are examples in tests drive, I guess if I manage to have rust bindings, it should be enough for me. |
That's great, please let me know if you need anything. The Visual Studio 2017 solution contains 4 projects:
Unfortunately, I have no experience with Rust but if you can make a wrapper for C++ code, then you should have no problems. It would be nice if you could have your work in a GitHub repository, so that we can learn from the process! |
First thanks for the library!
What is the recommended approach to write large datasets (e.g. 20+ GB csv files). Is there any way to stream reading / writing ?
I have a hard time finding documentation on how to use it. The only one I found uses data frames. I am not an expert on R but I think it is in memory only.
Also I would ideally like to use it in a rust program, which means I'll probably need to do a rust binding for the required parts. Happy to share it if you want!
The text was updated successfully, but these errors were encountered: