Training & Optimization
Tensor Parallelism
Splitting individual layers/tensors across devices for very large models.
This concept is essential for understanding training & optimization and forms a key part of modern AI systems.
Related Concepts
- Model Parallelism
- Distributed Training
- Large Models
Tags
training-optimization model-parallelism distributed-training large-models