ONNX Runtime
Cross-platform ML inference engine supporting ONNX models on diverse hardware
ONNX Runtime
Cross-platform ML inference engine supporting ONNX models on diverse hardware
ONNX Runtime is an open-source cross-platform inference and training acceleration library from Microsoft that runs models in the Open Neural Network Exchange (ONNX) format on CPUs, GPUs, and specialized AI accelerators. It includes hardware-specific execution providers that optimize performance for Intel, NVIDIA, AMD, ARM, and other processors. Developers building ML applications across platforms use ONNX Runtime to deploy trained models from PyTorch, TensorFlow, or scikit-learn in production environments including web, mobile, cloud, and edge without framework-specific dependencies.
Key Features
- ✓Cross-platform
- ✓Hardware acceleration
- ✓ONNX format
- ✓Multiple execution providers
- ✓Edge and cloud
Quick Info
- Category
- AI Infrastructure
- Pricing
- Free
More AI Infrastructure Tools
Inferless
AI InfrastructureServerless AI model deployment platform with GPU auto-scaling and cold start optimization
Colossal AI
AI InfrastructureOpen-source system for efficient large-scale AI model training and fine-tuning
Neural Magic
AI InfrastructureSoftware-defined AI inference engine that runs LLMs at GPU speed on CPUs
Weaviate Cloud
AI InfrastructureFully managed cloud service for the Weaviate open-source vector database