-
Notifications
You must be signed in to change notification settings - Fork 25
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
DaskGeoDataFrame parquet write error - Series object has no attribute total_bounds #138
Comments
If there is only one Dataframe partition saving works fine - if there is > 1 partition, this error is returned. |
I would guess that this was implemented with |
Thanks for that idea @hoxbro. I downgraded dask to 2020 but it returns the same error. So far in looking into the issue I found that any call to |
Did you try to set the parquet backend to fastparquet? |
I did try fastparquet (same error). However, I don't think it's related to that or to saving directly. Something happens with pack_partitions that causes and future calls to the geometry.total_bounds property to fail. It's failing at save because to_parquet makes calls to that property. |
I found a trigger condition for the error - it occurs when one or more longitudes are negative. I attached a simple notebook that reproduces the error. If you change the negative longitude to positive the error is resolved. Not sure where to look in the code to patch this. Thanks! |
Hi - I'm running into an error when trying to write a
DaskGeoDataFrame
. I'm following the basic pattern here (see also) but using a smaller sample of a point dataset. Everything seems to run as expected until trying to write out the packed file and I encounter the error below.ALL software version info
pyarrow =15.0.0
spatialpandas=0.4.10
pandas=2.1.1
dask=2024.2.0
python=3.9.16
The text was updated successfully, but these errors were encountered: