Orloj – agent infrastructure as code
FrameworkFreeHey HN, we're Jon and Kristiane, and we're building Orloj (https://orloj.dev), an open-source orchestration runtime for multi-agent AI systems. You define agents, tools, policies, and workflows in declarative YAML manifests, and Orloj handles scheduling, execution, governance, an
Capabilities13 decomposed
yaml-based agent workflow definition
Medium confidenceEnables declarative specification of multi-step agent workflows using YAML configuration files, where each step defines tool invocations, conditional branching, and state transitions. The framework parses YAML schemas into executable agent graphs, supporting sequential execution, parallel branches, and loop constructs without requiring imperative code. This approach treats agent orchestration as infrastructure-as-code, enabling version control, code review, and GitOps-style deployment patterns.
Applies GitOps and infrastructure-as-code patterns to agent workflows, enabling version-controlled, peer-reviewed agent configurations rather than treating agent logic as ephemeral code
Differs from LangChain/LlamaIndex by prioritizing declarative YAML configuration over imperative Python chains, enabling non-engineers to modify agent behavior and supporting GitOps deployment patterns
tool/function schema registration and binding
Medium confidenceProvides a registry system for declaring available tools and functions that agents can invoke, with automatic schema generation and validation. The framework accepts tool definitions (name, description, parameters, return types) and generates JSON schemas compatible with LLM function-calling APIs. Tools are bound to agent steps via references in YAML, with the framework handling parameter marshaling, type coercion, and error handling between the LLM output and actual function invocation.
Centralizes tool definitions in a declarative registry that generates LLM-compatible schemas automatically, reducing the gap between tool implementation and agent configuration
More structured than LangChain's tool decorators by enforcing schema validation upfront; simpler than Anthropic's native function-calling by abstracting multi-provider differences
human-in-the-loop approval workflows
Medium confidenceEnables agents to pause execution and request human approval before proceeding with critical actions. The framework provides mechanisms for sending approval requests (via email, Slack, webhooks, or UI), waiting for human decisions, and resuming execution based on approval/rejection. Approval workflows can be conditional (only require approval for certain actions) and can include context (reason for action, impact assessment) to help humans make informed decisions.
Provides declarative human-in-the-loop workflows in YAML, enabling approval gates without custom code
More integrated than manual approval processes by automating notification and decision tracking; simpler than building custom approval systems
parallel step execution and fan-out/fan-in patterns
Medium confidenceEnables agents to execute multiple steps in parallel and combine results, supporting fan-out (one step spawns multiple parallel tasks) and fan-in (multiple parallel tasks converge into one) patterns. The framework manages parallel execution, collects results, and provides mechanisms for combining results (merge, aggregate, select best). This is useful for tasks like querying multiple APIs in parallel or running multiple agent branches simultaneously.
Provides declarative parallel execution patterns in YAML, enabling fan-out/fan-in workflows without manual concurrency management
Simpler than building custom parallel orchestration; more efficient than sequential execution for I/O-bound operations
webhook and event-driven agent triggering
Medium confidenceEnables agents to be triggered by external events (webhooks, message queues, scheduled jobs) rather than only by direct API calls. The framework listens for incoming webhooks, parses event payloads, and automatically invokes agents with event data as input. This enables event-driven architectures where agents react to external system changes (e.g., new customer signup, payment received, alert triggered).
Provides declarative webhook and event-driven triggering in YAML, enabling agents to react to external events without custom code
More integrated than manual webhook handling; simpler than building custom event routing systems
multi-llm provider abstraction and routing
Medium confidenceAbstracts differences between LLM providers (OpenAI, Anthropic, local models, etc.) behind a unified interface, allowing agents to switch providers or route requests based on cost, latency, or capability requirements. The framework handles provider-specific API differences (function-calling formats, token counting, streaming behavior) transparently, enabling YAML configurations to specify provider preferences without embedding provider-specific code.
Provides declarative provider routing and fallback policies in YAML, enabling cost and latency optimization without code changes, rather than hardcoding provider selection
More flexible than LangChain's LLMChain by supporting dynamic provider routing; simpler than building custom provider adapters by handling API differences automatically
agent execution and state management
Medium confidenceExecutes agent workflows defined in YAML by maintaining execution state across steps, managing context windows, and handling step-to-step data flow. The framework tracks agent state (current step, variables, tool results, LLM responses) and provides mechanisms for passing data between steps, persisting state across invocations, and recovering from failures. Execution can be synchronous (blocking until completion) or asynchronous (with webhooks/callbacks for result notification).
Treats agent execution as a first-class workflow primitive with explicit state management and recovery semantics, rather than treating it as a simple function call
More robust than LangChain's basic chain execution by providing built-in state persistence and recovery; simpler than Temporal/Durable Functions by focusing specifically on agent workflows
conditional branching and loop constructs in workflows
Medium confidenceEnables agents to make decisions and repeat actions based on conditions evaluated at runtime. The framework supports if/else branching (based on LLM outputs, tool results, or variables), loops (repeat steps until condition met), and switch statements (route to different steps based on categorical decisions). Conditions are expressed in a simple DSL or as references to step outputs, allowing complex workflows without imperative code.
Provides declarative control flow primitives in YAML that avoid imperative code while supporting complex agent decision-making patterns
More readable than imperative Python chains for simple conditionals; less powerful than full programming languages but sufficient for most agent workflows
gitops-based agent deployment and versioning
Medium confidenceIntegrates with Git repositories to enable GitOps-style deployment of agent configurations. Changes to YAML workflow files trigger automatic deployments, with Git history serving as the audit trail for all agent behavior changes. The framework supports rollback to previous versions, branch-based testing (deploy from feature branches to staging), and pull request workflows where agent changes are reviewed before merging.
Applies GitOps principles to agent management, treating agent configurations as code with full version control, peer review, and automated deployment
Enables agent governance that REST API-based agent platforms lack; simpler than building custom deployment pipelines by providing GitOps integration out-of-the-box
agent monitoring and execution observability
Medium confidenceProvides built-in logging, metrics, and tracing for agent executions, capturing step-by-step execution details, LLM API calls, tool invocations, and performance metrics. The framework emits structured logs that can be sent to external observability platforms (Datadog, New Relic, etc.) and provides dashboards or CLI tools for viewing execution traces. Metrics include latency per step, token usage, cost, and error rates.
Provides first-class observability for agent workflows with automatic metric collection and structured logging, rather than requiring manual instrumentation
More comprehensive than LangChain's basic logging by capturing cost and performance metrics automatically; simpler than building custom observability by providing built-in integrations
prompt templating and variable interpolation
Medium confidenceSupports templating of LLM prompts with variables that are substituted at runtime based on agent context, tool results, or user input. The framework uses a simple template syntax (e.g., {{variable_name}}) to embed variables in prompts, with support for filters (e.g., {{variable | uppercase}}) and conditional sections (e.g., {{#if condition}}...{{/if}}). This enables dynamic prompt generation without string concatenation in code.
Provides declarative prompt templating in YAML, enabling non-engineers to modify prompts without code changes
Simpler than LangChain's PromptTemplate for basic use cases; less powerful than full template engines but sufficient for agent workflows
error handling and retry policies
Medium confidenceProvides declarative error handling and retry mechanisms for agent steps, allowing configuration of retry counts, backoff strategies (exponential, linear, fixed), and fallback actions. When a step fails (tool error, LLM timeout, etc.), the framework automatically retries according to the configured policy or executes a fallback step. Error conditions can be caught and handled without terminating the entire workflow.
Provides declarative retry and error handling policies in YAML, eliminating the need for manual error handling code in agent logic
More flexible than simple try-catch by supporting configurable backoff strategies; simpler than circuit breaker patterns by focusing on individual step resilience
agent context and memory management
Medium confidenceManages context windows and memory for agents, tracking conversation history, tool results, and intermediate outputs across steps. The framework provides mechanisms for summarizing or pruning old context to stay within LLM token limits, storing long-term memory in external systems (databases, vector stores), and retrieving relevant context for each step. Context can be scoped to individual conversations, users, or global agent state.
Provides declarative context management policies in YAML, enabling automatic context trimming and memory management without manual code
More integrated than LangChain's memory classes by providing automatic context summarization; simpler than building custom memory systems
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with Orloj – agent infrastructure as code, ranked by overlap. Discovered automatically through the match graph.
GPTSwarm
Language Agents as Optimizable Graphs
Build agents via YAML with Prolog validation and 110 built-in tools
I'm one of the creators of The Edge Agent (TEA). We built this because we needed a way to deploy agents that was verifiable and robust enough for production/edge cases, moving away from loose scripts.The architecture aims to solve critical gaps in deterministic orchestration identified by
PraisonAI
A framework for building multi-agent AI systems with workflows, tool integrations, and memory. #opensource
ChatDev
Communicative agents for software development
deepagents
Agent harness built with LangChain and LangGraph. Equipped with a planning tool, a filesystem backend, and the ability to spawn subagents - well-equipped to handle complex agentic tasks.
ms-agent
MS-Agent: a lightweight framework to empower agentic execution of complex tasks
Best For
- ✓DevOps teams managing multiple agent deployments
- ✓Organizations requiring audit trails and change management for agent behavior
- ✓Teams without dedicated ML engineers who need to configure agents declaratively
- ✓Teams building agents that integrate with multiple external APIs and services
- ✓Organizations standardizing on a tool registry for agent reuse across projects
- ✓Developers who want automatic schema generation to reduce boilerplate
- ✓Regulated industries requiring human oversight of automated decisions
- ✓High-stakes applications where agent errors have significant consequences
Known Limitations
- ⚠Complex conditional logic may become verbose in YAML; deeply nested workflows reduce readability
- ⚠Limited to predefined step types and patterns defined in the framework schema
- ⚠Debugging YAML-defined workflows requires framework-specific tooling; stack traces may be opaque
- ⚠Tool discovery is static (defined at configuration time); dynamic tool registration at runtime may not be supported
- ⚠Complex parameter types (nested objects, unions) may require manual schema refinement
- ⚠No built-in caching of tool schemas; repeated schema generation adds latency
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
About
Show HN: Orloj – agent infrastructure as code (YAML and GitOps)
Categories
Alternatives to Orloj – agent infrastructure as code
Search the Supabase docs for up-to-date guidance and troubleshoot errors quickly. Manage organizations, projects, databases, and Edge Functions, including migrations, SQL, logs, advisors, keys, and type generation, in one flow. Create and manage development branches to iterate safely, confirm costs
Compare →Are you the builder of Orloj – agent infrastructure as code?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →