-
Notifications
You must be signed in to change notification settings - Fork 77
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Slow when writing to Zarr #29
Comments
@aolt first guess the You can use '{path}' in
Also chunking on latitude / longitude is not very useful with GRIB files in general, avoid it if you can. |
Thanks for the trick with |
@aolt improving performance, especially when using dask, is in our todo list, but not very high as long as we work on stabilising the API, sorry. |
I opened #33 for general dask performance issue, I leave this issue open for now, but my guess is that there is nothing specific to Zarr. |
thanks @alexamici, I think this is specific to cfgrib. I get much higher performance with reading NetCDF with xarray/scipy and writing into Zarr. |
I close the issue as it looks like a generic disk performance issue. |
I am trying to convert a multiple files grib files Xarray dataset into Zarr. Reading these files is relatively quick, but writing to Zarr is going very slow. What I am doing:
It writes about 1MB per 10 seconds and it is using 100% CPU. There is a lot performance capacity on the disk, so I assume it should be about how it reads grib files.
Each file is about 1.5GB and I have about 216 files to write as Zarr
The text was updated successfully, but these errors were encountered: