Skip to content

Commit 231533c

Browse files
committed
shuffle -> distributed_shuffle
1 parent d467bc6 commit 231533c

File tree

2 files changed

+15
-15
lines changed

2 files changed

+15
-15
lines changed

xarray/core/groupby.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -682,7 +682,7 @@ def sizes(self) -> Mapping[Hashable, int]:
682682
self._sizes = self._obj.isel({self._group_dim: index}).sizes
683683
return self._sizes
684684

685-
def shuffle(self, chunks: T_Chunks = None):
685+
def distributed_shuffle(self, chunks: T_Chunks = None):
686686
"""
687687
Sort or "shuffle" the underlying object.
688688

xarray/tests/test_groupby.py

Lines changed: 14 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -672,7 +672,7 @@ def test_groupby_drops_nans(shuffle: bool, chunk: Literal[False] | dict) -> None
672672
ds["variable"] = ds["variable"].chunk(chunk)
673673
grouped = ds.groupby(ds.id)
674674
if shuffle:
675-
grouped = grouped.shuffle()
675+
grouped = grouped.distributed_shuffle()
676676

677677
# non reduction operation
678678
expected1 = ds.copy()
@@ -1418,7 +1418,7 @@ def test_groupby_reductions(
14181418
with raise_if_dask_computes():
14191419
grouped = array.groupby("abc")
14201420
if shuffle:
1421-
grouped = grouped.shuffle()
1421+
grouped = grouped.distributed_shuffle()
14221422

14231423
with xr.set_options(use_flox=use_flox):
14241424
actual = getattr(grouped, method)(dim="y")
@@ -1689,11 +1689,11 @@ def test_groupby_bins(
16891689
gb = array.groupby_bins("dim_0", bins=bins, **cut_kwargs)
16901690
actual = gb.sum()
16911691
assert_identical(expected, actual)
1692-
assert_identical(expected, gb.shuffle().sum())
1692+
assert_identical(expected, gb.distributed_shuffle().sum())
16931693

16941694
actual = gb.map(lambda x: x.sum())
16951695
assert_identical(expected, actual)
1696-
assert_identical(expected, gb.shuffle().map(lambda x: x.sum()))
1696+
assert_identical(expected, gb.distributed_shuffle().map(lambda x: x.sum()))
16971697

16981698
# make sure original array dims are unchanged
16991699
assert len(array.dim_0) == 4
@@ -1880,14 +1880,14 @@ def resample_as_pandas(array, *args, **kwargs):
18801880
actual = rs.mean()
18811881
expected = resample_as_pandas(array, resample_freq)
18821882
assert_identical(expected, actual)
1883-
assert_identical(expected, rs.shuffle().mean())
1883+
assert_identical(expected, rs.distributed_shuffle().mean())
18841884

18851885
assert_identical(expected, rs.reduce(np.mean))
1886-
assert_identical(expected, rs.shuffle().reduce(np.mean))
1886+
assert_identical(expected, rs.distributed_shuffle().reduce(np.mean))
18871887

18881888
rs = array.resample(time="24h", closed="right")
18891889
actual = rs.mean()
1890-
shuffled = rs.shuffle().mean()
1890+
shuffled = rs.distributed_shuffle().mean()
18911891
expected = resample_as_pandas(array, "24h", closed="right")
18921892
assert_identical(expected, actual)
18931893
assert_identical(expected, shuffled)
@@ -2832,7 +2832,7 @@ def test_multiple_groupers(use_flox: bool, shuffle: bool) -> None:
28322832

28332833
gb = da.groupby(labels1=UniqueGrouper(), labels2=UniqueGrouper())
28342834
if shuffle:
2835-
gb = gb.shuffle()
2835+
gb = gb.distributed_shuffle()
28362836
repr(gb)
28372837

28382838
expected = DataArray(
@@ -2853,7 +2853,7 @@ def test_multiple_groupers(use_flox: bool, shuffle: bool) -> None:
28532853
square = DataArray(np.arange(16).reshape(4, 4), coords=coords, dims=["x", "y"])
28542854
gb = square.groupby(a=UniqueGrouper(), b=UniqueGrouper())
28552855
if shuffle:
2856-
gb = gb.shuffle()
2856+
gb = gb.distributed_shuffle()
28572857
repr(gb)
28582858
with xr.set_options(use_flox=use_flox):
28592859
actual = gb.mean()
@@ -2878,14 +2878,14 @@ def test_multiple_groupers(use_flox: bool, shuffle: bool) -> None:
28782878
)
28792879
gb = b.groupby(x=UniqueGrouper(), y=UniqueGrouper())
28802880
if shuffle:
2881-
gb = gb.shuffle()
2881+
gb = gb.distributed_shuffle()
28822882
repr(gb)
28832883
with xr.set_options(use_flox=use_flox):
28842884
assert_identical(gb.mean("z"), b.mean("z"))
28852885

28862886
gb = b.groupby(x=UniqueGrouper(), xy=UniqueGrouper())
28872887
if shuffle:
2888-
gb = gb.shuffle()
2888+
gb = gb.distributed_shuffle()
28892889
repr(gb)
28902890
with xr.set_options(use_flox=use_flox):
28912891
actual = gb.mean()
@@ -2939,7 +2939,7 @@ def test_multiple_groupers_mixed(use_flox: bool, shuffle: bool) -> None:
29392939
)
29402940
gb = ds.groupby(x=BinGrouper(bins=[5, 15, 25]), letters=UniqueGrouper())
29412941
if shuffle:
2942-
gb = gb.shuffle()
2942+
gb = gb.distributed_shuffle()
29432943
expected_data = np.array(
29442944
[
29452945
[[0.0, np.nan], [np.nan, 3.0]],
@@ -3176,12 +3176,12 @@ def test_shuffle_by_simple() -> None:
31763176
data=dask.array.from_array([1, 2, 3, 4, 5, 6], chunks=2),
31773177
coords={"label": ("x", "a b c a b c".split(" "))},
31783178
)
3179-
actual = da.shuffle_by(label=UniqueGrouper())
3179+
actual = da.distributed_shuffle_by(label=UniqueGrouper())
31803180
expected = da.isel(x=[0, 3, 1, 4, 2, 5])
31813181
assert_identical(actual, expected)
31823182

31833183
with pytest.raises(ValueError):
3184-
da.chunk(x=2, eagerly_load_group=False).shuffle_by("label")
3184+
da.chunk(x=2, eagerly_load_group=False).distributed_shuffle_by("label")
31853185

31863186

31873187
@requires_dask

0 commit comments

Comments
 (0)