KServe
Kubernetes-native model serving platform for ML inference at scale
KServe is a highly scalable Kubernetes-native model serving platform that provides serverless inference for ML frameworks including TensorFlow, PyTorch, scikit-learn, and custom models. It supports canary deployments, auto-scaling to zero, transformer pipelines, and multi-model serving, making it suitable for production ML systems that require operational sophistication. Platform engineering teams, MLOps practitioners, and cloud-native enterprises use KServe to manage model serving infrastructure on Kubernetes with the same reliability standards applied to other production services.
Key Features
- ✓Kubernetes-native
- ✓Auto-scaling
- ✓Multi-framework support
- ✓Canary deployments
- ✓Serverless inference
Quick Info
- Category
- AI Infrastructure
- Pricing
- Free
More AI Infrastructure Tools
Inferless
AI InfrastructureServerless AI model deployment platform with GPU auto-scaling and cold start optimization
Colossal AI
AI InfrastructureOpen-source system for efficient large-scale AI model training and fine-tuning
Neural Magic
AI InfrastructureSoftware-defined AI inference engine that runs LLMs at GPU speed on CPUs
Weaviate Cloud
AI InfrastructureFully managed cloud service for the Weaviate open-source vector database