Skip to content

Commit 066fd04

Browse files
authored
correct TP implementation resources (huggingface#13248)
fix a few implementation links
1 parent 4d10474 commit 066fd04

File tree

1 file changed

+5
-2
lines changed

1 file changed

+5
-2
lines changed

docs/source/parallelism.md

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -220,9 +220,12 @@ Special considerations: TP requires very fast network, and therefore it's not ad
220220
This section is based on the original much more [detailed TP overview](https://github.com/huggingface/transformers/issues/10321#issuecomment-783543530).
221221
by [@anton-l](https://github.com/anton-l).
222222

223-
Implementations:
223+
Alternative names:
224224
- DeepSpeed calls it [tensor slicing](https://www.deepspeed.ai/features/#model-parallelism)
225-
- [Megatron-LM](https://github.com/NVIDIA/Megatron-LM) has an internal implementation.
225+
226+
Implementations:
227+
- [Megatron-LM](https://github.com/NVIDIA/Megatron-LM) has an internal implementation, as it's very model-specific
228+
- [parallelformers](https://github.com/tunib-ai/parallelformers) (only inference at the moment)
226229

227230
🤗 Transformers status:
228231
- core: not yet implemented in the core

0 commit comments

Comments
 (0)