-
Notifications
You must be signed in to change notification settings - Fork 25
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
DaskGeoDataFrame parquet write error - Series object has no attribute total_bounds #138
Comments
If there is only one Dataframe partition saving works fine - if there is > 1 partition, this error is returned. |
I would guess that this was implemented with |
Thanks for that idea @hoxbro. I downgraded dask to 2020 but it returns the same error. So far in looking into the issue I found that any call to |
Did you try to set the parquet backend to fastparquet? |
I did try fastparquet (same error). However, I don't think it's related to that or to saving directly. Something happens with pack_partitions that causes and future calls to the geometry.total_bounds property to fail. It's failing at save because to_parquet makes calls to that property. |
I found a trigger condition for the error - it occurs when one or more longitudes are negative. I attached a simple notebook that reproduces the error. If you change the negative longitude to positive the error is resolved. Not sure where to look in the code to patch this. Thanks! |
Hi - I'm running into an error when trying to write a
DaskGeoDataFrame
. I'm following the basic pattern here (see also) but using a smaller sample of a point dataset. Everything seems to run as expected until trying to write out the packed file and I encounter the error below.ALL software version info
pyarrow =15.0.0
spatialpandas=0.4.10
pandas=2.1.1
dask=2024.2.0
python=3.9.16
The text was updated successfully, but these errors were encountered: