Microsoft DeepSpeed
Deep learning optimization library for training and serving massive models efficiently
Microsoft DeepSpeed
Deep learning optimization library for training and serving massive models efficiently
DeepSpeed is an open-source deep learning optimization library from Microsoft Research that enables training and inference of extremely large AI models (hundreds of billions of parameters) more efficiently through techniques including ZeRO memory optimization, pipeline parallelism, and kernel fusion. DeepSpeed dramatically reduces the hardware required to train large models and has been used to train models including Megatron-Turing NLG and BLOOM. AI research labs, enterprises training custom large models, and teams scaling up ML training use DeepSpeed to overcome GPU memory and compute constraints.
Key Features
- ✓ZeRO memory optimization
- ✓Pipeline parallelism
- ✓Large model training
- ✓Inference optimization
- ✓Open-source
Quick Info
- Category
- AI Infrastructure & MLOps
- Pricing
- Free
More AI Infrastructure & MLOps Tools
Dstack
AI Infrastructure & MLOpsOpen-source cloud-agnostic platform for AI/ML workload orchestration
Tigris Data
AI Infrastructure & MLOpsAI-native object storage with built-in vector search and S3 compatibility
Superlinked
AI Infrastructure & MLOpsVector compute framework that helps ML engineers build retrieval systems by combining multiple data types a…
Qdrant Cloud
AI Infrastructure & MLOpsManaged vector database cloud service offering high-performance similarity search with filtering, payload i…