mirror of
https://github.com/open-mmlab/mmengine.git
synced 2025-06-03 21:54:44 +08:00
[Docs] Fix nnodes in the doc of ddp training (#1462)
This commit is contained in:
parent
1398e4200e
commit
369f15e27a
@ -26,7 +26,7 @@ On the first machine:
|
|||||||
|
|
||||||
```bash
|
```bash
|
||||||
python -m torch.distributed.launch \
|
python -m torch.distributed.launch \
|
||||||
--nnodes 8 \
|
--nnodes 2 \
|
||||||
--node_rank 0 \
|
--node_rank 0 \
|
||||||
--master_addr 127.0.0.1 \
|
--master_addr 127.0.0.1 \
|
||||||
--master_port 29500 \
|
--master_port 29500 \
|
||||||
@ -38,9 +38,9 @@ On the second machine:
|
|||||||
|
|
||||||
```bash
|
```bash
|
||||||
python -m torch.distributed.launch \
|
python -m torch.distributed.launch \
|
||||||
--nnodes 8 \
|
--nnodes 2 \
|
||||||
--node_rank 1 \
|
--node_rank 1 \
|
||||||
--master_addr 127.0.0.1 \
|
--master_addr "ip_of_the_first_machine" \
|
||||||
--master_port 29500 \
|
--master_port 29500 \
|
||||||
--nproc_per_node=8 \
|
--nproc_per_node=8 \
|
||||||
examples/distributed_training.py --launcher pytorch
|
examples/distributed_training.py --launcher pytorch
|
||||||
|
@ -26,7 +26,7 @@ CUDA_VISIBLE_DEVICES=0,3 python -m torch.distributed.launch --nproc_per_node=2 e
|
|||||||
|
|
||||||
```bash
|
```bash
|
||||||
python -m torch.distributed.launch \
|
python -m torch.distributed.launch \
|
||||||
--nnodes 8 \
|
--nnodes 2 \
|
||||||
--node_rank 0 \
|
--node_rank 0 \
|
||||||
--master_addr 127.0.0.1 \
|
--master_addr 127.0.0.1 \
|
||||||
--master_port 29500 \
|
--master_port 29500 \
|
||||||
@ -38,9 +38,9 @@ python -m torch.distributed.launch \
|
|||||||
|
|
||||||
```bash
|
```bash
|
||||||
python -m torch.distributed.launch \
|
python -m torch.distributed.launch \
|
||||||
--nnodes 8 \
|
--nnodes 2 \
|
||||||
--node_rank 1 \
|
--node_rank 1 \
|
||||||
--master_addr 127.0.0.1 \
|
--master_addr "ip_of_the_first_machine" \
|
||||||
--master_port 29500 \
|
--master_port 29500 \
|
||||||
--nproc_per_node=8 \
|
--nproc_per_node=8 \
|
||||||
examples/distributed_training.py --launcher pytorch
|
examples/distributed_training.py --launcher pytorch
|
||||||
|
Loading…
x
Reference in New Issue
Block a user