[DDP Note] Remove refs to RoundRobin PG until we officially support it (#40380)
Summary:
Removes line mentioning `ProcessGroupRoundRobin` since we don't intend it to be used as a public API just yet. We can add this back when we officially support the API
Pull Request resolved: https://github.com/pytorch/pytorch/pull/40380
Differential Revision: D22165556
Pulled By: rohan-varma
fbshipit-source-id: 24d0477d881dc74f2ff579de61dfd1ced2b09e75
diff --git a/docs/source/notes/ddp.rst b/docs/source/notes/ddp.rst
index 5634482..8b54098 100644
--- a/docs/source/notes/ddp.rst
+++ b/docs/source/notes/ddp.rst
@@ -145,14 +145,11 @@
- `ProcessGroup.hpp <https://github.com/pytorch/pytorch/blob/v1.4.0/torch/lib/c10d/ProcessGroup.hpp>`__:
contains the abstract API of all process group implementations. The ``c10d``
- library provides 4 implementations out of the box, namely,
- `ProcessGroupGloo`, `ProcessGroupNCCL`, `ProcessGroupMPI`, and
- `ProcessGroupRoundRobin`, where `ProcessGroupRoundRobin` is a composition of
- multiple process group instances and launches collective communications in a
- round-robin manner. ``DistributedDataParallel`` uses
- ``ProcessGroup::broadcast()`` to send model states from the process with rank
- 0 to others during initialization and ``ProcessGroup::allreduce()`` to sum
- gradients.
+ library provides 3 implementations out of the box, namely,
+ `ProcessGroupGloo`, `ProcessGroupNCCL`, and `ProcessGroupMPI`.
+ ``DistributedDataParallel`` uses ``ProcessGroup::broadcast()`` to send
+ model states from the process with rank 0 to others during initialization
+ and ``ProcessGroup::allreduce()`` to sum gradients.
- `Store.hpp <https://github.com/pytorch/pytorch/blob/v1.4.0/torch/lib/c10d/Store.hpp>`__: