[Docs] Fix nnodes in the doc of ddp training (#1462)

This commit is contained in:
XiwuChen 2024-01-02 10:42:58 +08:00 committed by GitHub
parent 1398e4200e
commit 369f15e27a
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
2 changed files with 6 additions and 6 deletions

View File

@ -26,7 +26,7 @@ On the first machine:
```bash
python -m torch.distributed.launch \
--nnodes 8 \
--nnodes 2 \
--node_rank 0 \
--master_addr 127.0.0.1 \
--master_port 29500 \
@ -38,9 +38,9 @@ On the second machine:
```bash
python -m torch.distributed.launch \
--nnodes 8 \
--nnodes 2 \
--node_rank 1 \
--master_addr 127.0.0.1 \
--master_addr "ip_of_the_first_machine" \
--master_port 29500 \
--nproc_per_node=8 \
examples/distributed_training.py --launcher pytorch

View File

@ -26,7 +26,7 @@ CUDA_VISIBLE_DEVICES=0,3 python -m torch.distributed.launch --nproc_per_node=2 e
```bash
python -m torch.distributed.launch \
--nnodes 8 \
--nnodes 2 \
--node_rank 0 \
--master_addr 127.0.0.1 \
--master_port 29500 \
@ -38,9 +38,9 @@ python -m torch.distributed.launch \
```bash
python -m torch.distributed.launch \
--nnodes 8 \
--nnodes 2 \
--node_rank 1 \
--master_addr 127.0.0.1 \
--master_addr "ip_of_the_first_machine" \
--master_port 29500 \
--nproc_per_node=8 \
examples/distributed_training.py --launcher pytorch