-
Notifications
You must be signed in to change notification settings - Fork 84
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
Create a MultiZarr json file from netcdf files of unequal time length. #447
Comments
No, you unfortunately cannot "subchunk" the data that have chunk sizes > 1. The sole exception is completely uncompressed/encoded data, which I assume is not your situation. Explanation: |
Ok, thanks a lot, very much clear. Therefore I will proceed creating two different json files, one for the 12-step chunk and one for 1-step chunk, and then merging afterwards when opening them. in principle this could work! |
Hi there,
I am very new to
kerchunk
but I am trying to create a json file using zarr starting from a series of netcdf files, which might have unequal time (they can be 1 or 12). I do not want to replicate the data with zarr, for both storage and backward compatibility reasons.Here below a very basic example, I can attach the data but I think it is clear what is done here.
This fails with:
Browsing various issues in the repository (as this #430 (comment)) , it seems that this is due to a discussed limitation of Zarr that does not allow for unequal chunk sizes, which goes beyond
kerchunk
.However, I am wondering if there is a way to force the chunk when accessing the data so that for example if I set
chunks={"time": 1}
, as for example done with xarray, I should be able to still load the data.Thanks a lot for any hint!
The text was updated successfully, but these errors were encountered: