Saving parquet to AWS S3 with df.write_parquet() fails with FileNotFound #19930
Labels
bug
Something isn't working
needs triage
Awaiting prioritization by a maintainer
python
Related to Python Polars
Checks
Reproducible example
Log output
No response
Issue description
When passing a path-like object to write_parquet, I am getting a FileNotFoundError. This is almost the same as #14630, but here I'm trying to write to an s3 URI.
The same workaround works for S3 as well:
df.write_parquet("s3://nomadresearch-research-storage/alex/empty.parquet", use_pyarrow=True)
.Until this feature is supported natively in Rust, could we have a sentence in the documentation of write_parquet that when we pass a path-like
file
argument, we should useuse_pyarrow=True
for certain storage backends?Expected behavior
I would expect the write_parquet to create or overwrite the file at the given path. I recently changed my code because I was getting warnings that write_parquet didn't like it that I was passing it file-handles and that I should be passing path-like values, but this has broken my code.
Installed versions
Note I also tried with fsspec==2024.06.0
The text was updated successfully, but these errors were encountered: