via “vision-language model inference with multimodal input handling”
Run frontier LLMs and VLMs with day-0 model support across GPU, NPU, and CPU, with comprehensive runtime coverage for PC (Python/C++), mobile (Android & iOS), and Linux/IoT (Arm64 & x86 Docker). Supporting OpenAI GPT-OSS, IBM Granite-4, Qwen-3-VL, Gemma-3n, Ministral-3, and more.
Unique: VLM plugin architecture (runner/nexa-sdk/vlm.go) separates image encoding from text generation, allowing hardware-specific optimization of vision towers (GPU tensor cores for image embeddings) while text generation runs on NPU, maximizing throughput on heterogeneous hardware.
vs others: Only on-device VLM framework supporting NPU acceleration for vision encoding, whereas competitors (Ollama, LM Studio) run full VLM on single GPU, making it 3-5x more efficient on mobile/edge devices with heterogeneous compute.