-
Notifications
You must be signed in to change notification settings - Fork 6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to reduce memory usage? #74
Comments
Welcome back, @Kazuuk! The Python SBDF module already uses table slices as you describe (and I originally commented in spotfiresoftware/spotfire-sbdf-c#8); it's controlled by the
If you're willing to augment your code and install an additional package from PyPI, we can figure out which is the culprit. (We can't use the builtin
This is basically capturing the RSS (resident set size) memory usage of the Python process (and this is where the |
Thank you for your kind help :) Python Code is as follows, Output lines starts with SFPY is as follows, Case #2 (Parquet file size as 573MB, originally 2.1GB in CSV) Thank you again in advance :) |
Hello
I tried to 'export_data' with Parquet file, which source is the large(2GB) size CSV, to SBDF.
When I tried with C library, I used 'tableslice' as recommended and it works fine.
But in Python, I converted Parquet file to Pandas dataframe first and then convert Pandas dataframe to SBDF.
In the conversion that Pandas dataframe to SBDF consume a lot of memory.
Is there anyway to use 'tableslice' method in Python, or better way?
Thank you in advance.
The text was updated successfully, but these errors were encountered: