via “openai-compatible http server with function calling and streaming”
Run frontier LLMs and VLMs with day-0 model support across GPU, NPU, and CPU, with comprehensive runtime coverage for PC (Python/C++), mobile (Android & iOS), and Linux/IoT (Arm64 & x86 Docker). Supporting OpenAI GPT-OSS, IBM Granite-4, Qwen-3-VL, Gemma-3n, Ministral-3, and more.
Unique: Schema-based function registry (runner/server/service/) implements OpenAI and Anthropic function-calling protocols natively, allowing agents built for cloud APIs to execute local tools without adapter code. Middleware stack enables request/response transformation without modifying core inference logic.
vs others: Provides OpenAI API compatibility with function calling support, unlike Ollama which lacks structured tool calling, and unlike LM Studio which has no HTTP server at all, making it the only on-device framework that can replace cloud LLM APIs for agent workflows.