via “function calling with schema-based tool registry and multi-provider support”
Run frontier LLMs and VLMs with day-0 model support across GPU, NPU, and CPU, with comprehensive runtime coverage for PC (Python/C++), mobile (Android & iOS), and Linux/IoT (Arm64 & x86 Docker). Supporting OpenAI GPT-OSS, IBM Granite-4, Qwen-3-VL, Gemma-3n, Ministral-3, and more.
Unique: Schema-based function registry (runner/server/service/) implements both OpenAI and Anthropic function-calling protocols with unified interface, enabling agents built for cloud APIs to execute local tools without adapter code. Middleware stack enables request/response transformation without modifying core inference.
Supports both OpenAI and Anthropic function-calling protocols natively, whereas Ollama has no function calling support and LM Studio requires manual JSON parsing, making it the only on-device framework enabling true multi-provider agent compatibility.