diff --git a/python/cugraph/cugraph/dask/common/part_utils.py b/python/cugraph/cugraph/dask/common/part_utils.py index 0f5d46e83bc..bfa32bc49db 100644 --- a/python/cugraph/cugraph/dask/common/part_utils.py +++ b/python/cugraph/cugraph/dask/common/part_utils.py @@ -115,9 +115,8 @@ def persist_dask_df_equal_parts_per_worker( raise ValueError("return_type must be either 'dask_cudf.DataFrame' or 'dict'") ddf_keys = dask_df.to_delayed() - workers = client.scheduler_info()["workers"].keys() worker_to_rank = Comms.rank_to_worker(client) - # assure rank-worker mappings are in ascending order + # rank-worker mappings are in ascending order workers = dict(sorted(worker_to_rank.items())).values() ddf_keys_ls = _chunk_lst(ddf_keys, len(workers)) diff --git a/python/cugraph/cugraph/dask/comms/comms.py b/python/cugraph/cugraph/dask/comms/comms.py index 3897ab4c959..5499b13af03 100644 --- a/python/cugraph/cugraph/dask/comms/comms.py +++ b/python/cugraph/cugraph/dask/comms/comms.py @@ -269,7 +269,7 @@ def get_n_workers(sID=None, dask_worker=None): def rank_to_worker(client): """ - Return a mapping of dask workers to ranks. + Return a mapping of ranks to dask workers. """ workers = client.scheduler_info()["workers"].keys() worker_info = __instance.worker_info(workers)