Skip to content

Commit da03a9b

Browse files
committed
sync with #16860
1 parent 0b74488 commit da03a9b

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

docs/source/en/perf_hardware.mdx

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -17,15 +17,15 @@ limitations under the License.
1717

1818
# Custom hardware for training
1919

20-
The hardware you use to run model training and inference can have a big effect on performance. For a deep dive into GPUs make sure to check out Tim Dettmer's excellent [blog post](https://timdettmers.com/2020/09/07/which-gpu-for-deep-learning/).
20+
The hardware you use to run model training and inference can have a big effect on performance. For a deep dive into GPUs make sure to check out Tim Dettmer's excellent [blog post](https://timdettmers.com/2020/09/07/which-gpu-for-deep-learning/).
2121

2222
Let's have a look at some practical advice for GPU setups.
2323

2424
## GPU
2525
When you train bigger models you have essentially three options:
2626
- bigger GPUs
2727
- more GPUs
28-
- more CPU and NVMe (offloaded to by [DeepSpeed-Infinity](deepspeed#nvme-support))
28+
- more CPU and NVMe (offloaded to by [DeepSpeed-Infinity](main_classes/deepspeed#nvme-support))
2929

3030
Let's start at the case where you have a single GPU.
3131

@@ -147,4 +147,4 @@ rm -r /tmp/test-clm; CUDA_VISIBLE_DEVICES=0,1 NCCL_P2P_DISABLE=1 python -m torch
147147
```
148148

149149
Hardware: 2x TITAN RTX 24GB each + NVlink with 2 NVLinks (`NV2` in `nvidia-smi topo -m`)
150-
Software: `pytorch-1.8-to-be` + `cuda-11.0` / `transformers==4.3.0.dev0`
150+
Software: `pytorch-1.8-to-be` + `cuda-11.0` / `transformers==4.3.0.dev0`

0 commit comments

Comments
 (0)