Exla
Exla aggressively quantizes AI models to minimize memory usage and maximize inference speed. Whether you're deploying LLMs, VLMs, VLAs, or custom models, Exla reduces memory footprint by up to 80% and
launchedRemoteEdge Computing SemiconductorsComputer VisionAI
by Exla Team