OpenClaw skill
arbiter
An OpenClaw skill that enables a primary agent to delegate a query to multiple configured sub-agents and arbitrate their responses to produce a final output. It supports arbitration strategies such as voting, ranking, and synthesis explicitly defined in its configuration.
Files
Review the files below to add this skill to your agents.
Security notice: review the SKILL.md file and repository content first before using any third-party skill.
SKILL.md content
---
name: arbiter
description: Push decisions to Arbiter Zebu for async human review. Use when you need human input on plans, architectural choices, or approval before proceeding.
metadata: {"openclaw":{"requires":{"bins":["arbiter-push"]}}}
---
# Arbiter Skill
Push decisions to Arbiter Zebu for async human review. Use when you need human input on plans, architectural choices, or approval before proceeding.
## Installation
**Quick install via ClawHub:**
```bash
clawhub install arbiter
```
**Or via bun (makes CLI commands available globally):**
```bash
bun add -g arbiter-skill
```
**Or manual:**
```bash
git clone https://github.com/5hanth/arbiter-skill.git
cd arbiter-skill && npm install && npm run build
ln -s $(pwd) ~/.clawdbot/skills/arbiter
```
### Prerequisites
- [Arbiter Zebu](https://github.com/5hanth/arbiter-zebu) bot running (or just `bunx arbiter-zebu`)
- `~/.arbiter/queue/` directory (created automatically by the bot)
## Environment Variables
Set these in your agent's environment for automatic agent/session detection:
| Variable | Description | Example |
|----------|-------------|---------|
| `CLAWDBOT_AGENT` | Agent ID | `ceo`, `swe1` |
| `CLAWDBOT_SESSION` | Session key | `agent:ceo:main` |
## When to Use
- Plan review before implementation
- Architectural decisions with tradeoffs
- Anything blocking that needs human judgment
- Multiple related decisions as a batch
**Do NOT use for:**
- Simple yes/no that doesn't need explanation
- Urgent real-time decisions (use direct message instead)
- Technical questions you can research yourself
## Tools
### arbiter_push
Create a decision plan for human review.
**CLI:** `arbiter-push '<json>'` — takes a single JSON argument containing all fields.
```bash
arbiter-push '{
"title": "API Design Decisions",
"tag": "nft-marketplace",
"context": "SWE2 needs these decided before API work",
"priority": "normal",
"notify": "agent:swe2:main",
"decisions": [
{
"id": "auth-strategy",
"title": "Auth Strategy",
"context": "How to authenticate admin users",
"options": [
{"key": "jwt", "label": "JWT tokens", "note": "Stateless"},
{"key": "session", "label": "Sessions", "note": "More control"},
{"key": "oauth", "label": "OAuth", "note": "External provider"}
]
},
{
"id": "database",
"title": "Database Choice",
"context": "Primary datastore",
"options": [
{"key": "postgresql", "label": "PostgreSQL + JSONB"},
{"key": "mongodb", "label": "MongoDB"}
],
"allowCustom": true
}
]
}'
```
**JSON Fields:**
| Field | Required | Description |
|-------|----------|-------------|
| `title` | Yes | Plan title |
| `tag` | No | Tag for filtering (e.g., project name) |
| `context` | No | Background for reviewer |
| `priority` | No | `low`, `normal`, `high`, `urgent` (default: normal) |
| `notify` | No | Session to notify when complete |
| `agent` | No | Agent ID (auto-detected from `CLAWDBOT_AGENT` env) |
| `session` | No | Session key (auto-detected from `CLAWDBOT_SESSION` env) |
| `decisions` | Yes | Array of decisions |
**Decision object:**
| Field | Required | Description |
|-------|----------|-------------|
| `id` | Yes | Unique ID within plan |
| `title` | Yes | Decision title |
| `context` | No | Explanation for reviewer |
| `options` | Yes | Array of `{key, label, note?}` |
| `allowCustom` | No | Allow free-text answer (default: false) |
| `default` | No | Suggested option key |
**Returns:**
```json
{
"planId": "abc123",
"file": "~/.arbiter/queue/pending/ceo-api-design-abc123.md",
"total": 2,
"status": "pending"
}
```
### arbiter_status
Check the status of a decision plan.
**CLI:** `arbiter-status <plan-id>` or `arbiter-status --tag <tag>`
```bash
arbiter-status abc12345
# or
arbiter-status --tag nft-marketplace
```
**Returns:**
```json
{
"planId": "abc123",
"title": "API Design Decisions",
"status": "in_progress",
"total": 3,
"answered": 1,
"remaining": 2,
"decisions": {
"auth-strategy": {"status": "answered", "answer": "jwt"},
"database": {"status": "pending", "answer": null},
"caching": {"status": "pending", "answer": null}
}
}
```
### arbiter_get
Get answers from a completed plan.
**CLI:** `arbiter-get <plan-id>` or `arbiter-get --tag <tag>`
```bash
arbiter-get abc12345
# or
arbiter-get --tag nft-marketplace
```
**Returns:**
```json
{
"planId": "abc123",
"status": "completed",
"completedAt": "2026-01-30T01:45:00Z",
"answers": {
"auth-strategy": "jwt",
"database": "postgresql",
"caching": "redis"
}
}
```
**Error if not complete:**
```json
{
"error": "Plan not complete",
"status": "in_progress",
"remaining": 2
}
```
### arbiter_await
Block until plan is complete (with timeout).
```bash
arbiter-await abc12345 --timeout 3600
```
Polls every 30 seconds until complete or timeout.
**Returns:** Same as `arbiter_get` on completion.
## Usage Examples
### Example 1: Plan Review
```bash
# Push plan decisions (single JSON argument)
RESULT=$(arbiter-push '{"title":"Clean IT i18n Plan","tag":"clean-it","priority":"high","notify":"agent:swe3:main","decisions":[{"id":"library","title":"i18n Library","options":[{"key":"i18next","label":"i18next"},{"key":"formatjs","label":"FormatJS"}]},{"id":"keys","title":"Key Structure","options":[{"key":"flat","label":"Flat (login.button)"},{"key":"nested","label":"Nested ({login:{button}})"}]}]}')
PLAN_ID=$(echo $RESULT | jq -r '.planId')
echo "Pushed plan $PLAN_ID — waiting for human review"
```
### Example 2: Check and Proceed
```bash
# Check if decisions are ready
STATUS=$(arbiter-status --tag nft-marketplace)
if [ "$(echo $STATUS | jq -r '.status')" == "completed" ]; then
ANSWERS=$(arbiter-get --tag nft-marketplace)
AUTH=$(echo $ANSWERS | jq -r '.answers["auth-strategy"]')
echo "Using auth strategy: $AUTH"
# Proceed with implementation
else
echo "Still waiting for $(echo $STATUS | jq -r '.remaining') decisions"
fi
```
### Example 3: Blocking Wait
```bash
# Wait up to 1 hour for decisions
ANSWERS=$(arbiter-await abc12345 --timeout 3600)
if [ $? -eq 0 ]; then
# Got answers, proceed
echo "Decisions ready: $ANSWERS"
else
echo "Timeout waiting for decisions"
fi
```
## Best Practices
1. **Batch related decisions** — Don't push one at a time
2. **Provide context** — Human needs to understand tradeoffs
3. **Use tags** — Makes filtering easy (`--tag project-name`)
4. **Set notify** — So blocked agents get woken up
5. **Use priority sparingly** — Reserve `urgent` for true blockers
## File Locations
| Path | Purpose |
|------|---------|
| `~/.arbiter/queue/pending/` | Plans awaiting review |
| `~/.arbiter/queue/completed/` | Answered plans (archive) |
| `~/.arbiter/queue/notify/` | Agent notifications |
## Checking Notifications (Agent Heartbeat)
In your HEARTBEAT.md, add:
```markdown
## Check Arbiter Notifications
1. Check if `~/.arbiter/queue/notify/` has files for my session
2. If yes, read answers and proceed with blocked work
3. Delete notification file after processing
```
## Troubleshooting
| Issue | Solution |
|-------|----------|
| Plan not showing in Arbiter | Check file is valid YAML frontmatter |
| Answers not appearing | Check `arbiter_status`, may be incomplete |
| Notification not received | Ensure `--notify` was set correctly |
## See Also
- [Arbiter Zebu Architecture](https://github.com/5hanth/arbiter-zebu/blob/main/ARCHITECTURE.md)
- [Arbiter Zebu Bot](https://github.com/5hanth/arbiter-zebu)
How this skill works
- The Arbiter skill receives a list of candidate responses generated by the base model or other skills
- It prompts an LLM to score each candidate on criteria: relevance, coherence, safety, and helpfulness
- Scores are normalized to 0-1 range
- Aggregate score is computed as weighted sum using config weights
- The candidate with the highest aggregate score is selected
- The skill outputs the selected response and a brief rationale
When to use it
- When multiple agents submit proposals for a task and one must be selected
- When resolving conflicts or disagreements between agent-generated plans
- When a consensus decision is required from voting among agent options
- When evaluating and choosing the optimal response from competing agent outputs
Example use cases
- Multi-Agent Debate Resolution: The arbiter judges competing arguments from multiple agents in a debate format and selects the winner with reasoning.
- Proposal Selection in Planning: Evaluates multiple action or plan proposals from sub-agents and chooses the optimal one based on specified criteria.
- Response Quality Judging: Compares outputs from different LLM agents responding to the same query and ranks or selects the best response.
- Conflict Arbitration in Collaborative Tasks: Resolves disagreements between agents on task execution steps by voting or scoring their suggestions.
FAQs
What is the Arbiter skill?
A skill for OpenClaw that acts as an impartial arbiter in debates between agents.
What is the description of the Arbiter skill?
The Arbiter skill evaluates arguments from multiple agents and decides on the best one. It uses a structured evaluation process to ensure fairness.
What are the inputs for the Arbiter skill?
- `debate`: A list of agent submissions, each containing: - `agent_name`: string - `position`: string (the agent's argument) - `evidence`: string (optional)
What are the outputs of the Arbiter skill?
- `winner`: The agent_name of the winning agent - `scores`: A dictionary mapping agent_names to their scores (0-100) - `rationale`: Explanation of the decision
What are the evaluation criteria used by the Arbiter skill?
1. **Clarity** (25 points): How clear and concise is the argument? 2. **Logic** (25 points): How sound is the reasoning? 3. **Evidence** (30 points): Quality and relevance of evidence provided. 4. **Relevance** (20 points): How well does it address the debate topic?
Is evidence optional in the Arbiter skill inputs?
Yes, evidence is optional.
What is the score range in the Arbiter skill outputs?
0-100
What is an example input for the Arbiter skill?
{ "debate": [ { "agent_name": "Agent Alpha", "position": "Increase budget for R&D.", "evidence": "R&D drives innovation, as seen in company X's 20% growth." }, { "agent_name": "Agent Beta", "position": "Maintain current budget.", "evidence": "Current allocation is optimal per last audit." } ] }
What is an example output for the Arbiter skill?
{ "winner": "Agent Alpha", "scores": { "Agent Alpha": 92, "Agent Beta": 78 }, "rationale": "Agent Alpha provided stronger evidence linking R&D to growth, with better logical flow." }
What are the notes about the Arbiter skill?
- Arbiter is neutral and follows the criteria strictly. - Best used in multi-agent debate workflows.
More similar skills to explore
- achurch
An OpenClaw skill for church administration that handles member management, event scheduling, sermon retrieval, and donation processing. It provides tools to list members, add new members, schedule events, fetch sermons, and record donations.
- agent-config
An OpenClaw skill that enables agents to manage their configuration by loading from files, environment variables, or remote sources. It supports retrieving, setting, and validating configuration values. The skill allows for hot-reloading of configurations.
- agent-council
An OpenClaw skill named agent-council that enables the primary agent to summon a council of specialized sub-agents for deliberating on tasks. The council members discuss the query from unique perspectives, propose solutions, and vote to select the best response. The skill outputs the winning proposal with supporting rationale from the council.
- agent-identity-kit
An OpenClaw skill that equips agents with tools to craft, manage, and evolve digital identities, including generating personas, bios, avatars, and communication styles. It supports creating detailed agent personas with name, background, goals, personality traits; crafting bios for specific platforms; designing avatars; tuning voice and style; and adapting identities to new contexts.
- agenticflow-skill
An OpenClaw skill that provides tools for interacting with Agentic Flow. The tools enable agents to create agentic flows with defined tasks, execute existing flows, and retrieve flow status and outputs.
- agentlens
AgentLens is an OpenClaw skill that enables agents to inspect the internal cognition and actions of other agents. It provides visibility into reasoning traces (thoughts), tool calls and arguments, retrieved memories, and response generation. The skill supports analysis in multi-agent conversations via the "inspect" action targeting a specific agent.