Replies: 1 comment 3 replies
-
this is a long-standing issue by now (see e.g. #7079). Note that this only happens with a threading scheduler, so you can use |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi all,
I am using
xarray
anddask
at scale, and have been running into issues with occasional complete segmentation faults. As in, some times on the same data and code it will fail, and other times it succeeds.Occasionally, it will report more helpful information, like
NetCDF- HDF error
,malloc
,free pointer
, but other times no such luck.The same code succeeds every time when running without a local dask cluster, but when a dask cluster is present, failures occur.
I am looking for any help / tips on diagnosing and identifying the cause of these errors. It clearly has something to do with dask, its worker setup and netcdf, but further then that I am not sure.
Example Error
These errors are occurring when loading large > 50 GB datasets from disk, modifying the shape, and values, and saving again to disk.
All files are netcdf compliant, and are being loaded with
xr.open_mfdataset
and saved with.to_netcdf
.Result of `xr.show_versions()`
What I have tried
.compute(scheduler='single-threaded')
.compute(rerun_exceptions_locally=True)
Beta Was this translation helpful? Give feedback.
All reactions