MLC LLM
Universal LLM deployment solution for running models on any device including mobile
MLC LLM
Universal LLM deployment solution for running models on any device including mobile
MLC LLM (Machine Learning Compilation for LLMs) is an open-source framework developed by the MLC AI team that compiles and optimizes large language models to run natively on diverse hardware including laptops, phones, edge devices, and web browsers. It uses TVM-based compilation to generate hardware-specific optimized code, enabling models to run at reasonable speeds even on consumer devices without GPU cloud access. Developers building offline-capable apps, privacy-focused applications, and edge AI deployments use MLC LLM to bring LLMs to hardware that cloud-based APIs cannot reach.
Key Features
- ✓Mobile deployment
- ✓Browser support
- ✓Hardware compilation
- ✓Edge AI
- ✓Multiple backends
Quick Info
- Category
- AI Infrastructure
- Pricing
- Free
More AI Infrastructure Tools
Inferless
AI InfrastructureServerless AI model deployment platform with GPU auto-scaling and cold start optimization
Colossal AI
AI InfrastructureOpen-source system for efficient large-scale AI model training and fine-tuning
Neural Magic
AI InfrastructureSoftware-defined AI inference engine that runs LLMs at GPU speed on CPUs
Weaviate Cloud
AI InfrastructureFully managed cloud service for the Weaviate open-source vector database