Replies: 3 comments 6 replies
-
Hi @rpgoldman , thanks for the report. There are a couple of things we need to know better in order to help you. Can you post the output of |
Beta Was this translation helpful? Give feedback.
-
@fmaussion Here's what I can gather:
I'm not sure that did anything. When I look at the chunks, I see:
|
Beta Was this translation helpful? Give feedback.
-
Follow-up 2: I made a MultiIndex in my pandas data frame, dropping many of the columns for values of independent variables. When I did this, and tried to translate to a Dataset, I get an out of memory error. I did a little further work, and realized I needed to install But ... then
There aren't any date time structures in this dataset, so I'm not sure why that failed. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I'm having trouble with a large data set that I am able to manipulate in memory, but that causes out of memory errors when writing to disk with
to_netcdf()
. The docs provide a wealth of information about different options, but there's so much that I find myself lost. Would it be possible to provide a general guideline for what to try when this kind of problem happens? For example, there is discussion of chunking, but I don't know if that's likely to help in this problem or not.I considered the possibility of using Dask to break up the saving computation, but the best I could come up with was this:
and I get an out of memory error before I get to trying to force this computation. (I was going to try
d_cont.compute(scheduler='processes')
next.)The backtrace seems to indicate that dask might not even be used, despite
compute=False
:Note that this is a dataset that I can read and write with Pandas and CSV. Presumably the more complex Xarray indexing is causing my problem here?
Follow up: Thing I did not expect
The pandas CSV for this data set (before I converted it from Dataframe to Dataset) is 786 meg. The nc file is 15.4 gig and it's not finished writing. So clearly I do not know what I am doing here.
Beta Was this translation helpful? Give feedback.
All reactions