diff --git a/README.md b/README.md index ae99e6e0..8ba65879 100644 --- a/README.md +++ b/README.md @@ -101,13 +101,7 @@ install using `python setup.py` You can run the tests by navigating to the `/xBOUT/` directory and entering `pytest`. - -It relies on two upstream additions to xarray -([first](https://github.com/pydata/xarray/pull/2482) & -[second](https://github.com/pydata/xarray/pull/2553) pull requests). -The first is merged, but the second isn't, so for now you need to clone -the branch of xarray containing the PR -[here](https://github.com/TomNicholas/xarray/tree/feature/nd_combine). +Requires xarray v0.12.2 or later. You will also need to install [dask](https://dask.org/), as described in the xarray documentation diff --git a/requirements.txt b/requirements.txt index e0a9ca3d..2df45747 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1,4 +1,4 @@ -git+https://github.com/TomNicholas/xarray@feature/nd_combine#egg=xarray +xarray >= 0.12.2 dask[array] >= 1.0.0 natsort >= 5.5.0 matplotlib >= 2.2 diff --git a/xbout/load.py b/xbout/load.py index 3f3a58bd..bc8e78b6 100644 --- a/xbout/load.py +++ b/xbout/load.py @@ -19,10 +19,10 @@ def _auto_open_mfboutdataset(datapath, chunks={}, info=True, keep_guards=True): _preprocess = partial(_trim, ghosts={'x': mxg, 'y': myg}) - ds = xarray.open_mfdataset(paths_grid, concat_dims=concat_dims, - data_vars='minimal', preprocess=_preprocess, - engine=filetype, chunks=chunks, - infer_order_from_coords=False) + ds = xarray.open_mfdataset(paths_grid, concat_dim=concat_dims, + combine='nested', data_vars='minimal', + preprocess=_preprocess, engine=filetype, + chunks=chunks) ds, metadata = _strip_metadata(ds)