Tools
Openclown
Multi-perspective AI task evaluation plugin for OpenClaw โ reply to any AI response with /clown to get expert critiques from multiple perspectives
Install
npm install
npm
README
# ๐ช OpenClown โ Multi-Perspective AI Task Evaluation
<p align="center">
<img src="docs/logo.png" width="300" alt="OpenClown logo"><br>
<em>Your AI did the work. Now let the circus review it.</em><br><br>
<a href="https://www.npmjs.com/package/openclown"><img src="https://img.shields.io/npm/v/openclown?style=for-the-badge" alt="npm version"></a>
<a href="LICENSE"><img src="https://img.shields.io/badge/License-MIT-blue.svg?style=for-the-badge" alt="MIT License"></a>
</p>
**OpenClown** is a plugin for [OpenClaw](https://github.com/openclaw/openclaw) that evaluates AI-completed tasks from multiple expert perspectives. A "circus" of specialized performers โ a philosopher, a security expert, a developer, and more โ independently critique what your AI assistant just did.
The goal: catch blind spots, surface risks, and improve AI output quality before you act on it.
## How it works
```
You ask OpenClaw to do something
โ
โผ
OpenClaw completes the task
โ
โผ
You reply to the response with /clown
(or just type /clown for the latest)
โ
โผ
โโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโโ
โ OpenClown Circus โ
โ โ
โ ๐ Identify target exchange โ
โ (reply content matching, โ
โ keyword, or latest) โ
โ ๐ Gather context โ
โ (target + up to 3 prior โ
โ exchanges for follow-ups โ
โ + tool calls & results) โ
โ โ โ
โ ๐ญ Philosopher โ questions assumptions
โ ๐ Security โ flags data exposure
โ ๐ป Developer โ spots better approaches
โ ... (12 performers available) โ
โโโโโโโโโโโโโโโโฌโโโโโโโโโโโโโโโโโโโโโโโ
โ
โผ
Formatted evaluation with
severity levels + actionable feedback
```
## Install
```bash
openclaw plugins install openclown
```
That's it. OpenClown automatically reuses the LLM provider and API key you configured during OpenClaw setup โ no extra configuration needed.
**Update to latest version:**
```bash
openclaw plugins update openclown
```
## Quick start
On mobile (WhatsApp, Telegram, Slack, Discord):
1. Ask OpenClaw a question and wait for the response
2. **Long-press** (or swipe) the AI response you want to evaluate
3. Tap **Reply**
4. Type `/clown` and send
That's it. OpenClown identifies the message you replied to and evaluates it with full context โ including the original question, any tool calls the AI made, and prior conversation history.
If you don't reply to a specific message, `/clown` automatically evaluates the **most recent** AI response.
| Channel | Reply + /clown | /clown (latest) | /clown \<keyword\> |
|---------|:-:|:-:|:-:|
| WhatsApp | โ
| โ
| โ
|
| Telegram | โ
| โ
| โ
|
| Discord | โ
| โ
| โ
|
| Slack | โ
| โ
| โ
|
| CLI | โ | โ
| โ
|
## `/clown` โ Evaluate
Reply to any AI response and type `/clown` to evaluate it. Or just send `/clown` to evaluate the latest response.
```
/clown # Evaluate latest AI response
/clown <keyword> # Search by keyword, e.g. /clown weather
```
Each evaluation shows results from multiple performers with severity levels:
- ๐ก **Insight** โ observations, suggestions, things to consider
- โ ๏ธ **Warning** โ potential risks or issues worth addressing
- ๐ด **Critical** โ serious problems that should be fixed
## `/clown encore` โ Improve
After an evaluation, reply with `/clown encore` to re-run the original task with the evaluation feedback applied. The AI re-answers your question, addressing the issues the performers found:
```
/clown encore # AI re-answers, addressing the issues found
```
## Usage examples
OpenClown works anywhere OpenClaw does โ WhatsApp, Telegram, Slack, Discord, or CLI.
### Ask a question, then evaluate
<img src="https://raw.githubusercontent.com/Bubbletea98/openclown/main/docs/screenshots/01-ask0-and-response.JPG" width="300" alt="User asks OpenClaw a question on WhatsApp">
### `/clown` โ multi-perspective evaluation
<img src="https://raw.githubusercontent.com/Bubbletea98/openclown/main/docs/screenshots/02-clown-evaluation-1.JPG" width="300" alt="OpenClown evaluation โ part 1">
<img src="https://raw.githubusercontent.com/Bubbletea98/openclown/main/docs/screenshots/02-clown-evaluation-2.JPG" width="300" alt="OpenClown evaluation โ part 2">
<img src="https://raw.githubusercontent.com/Bubbletea98/openclown/main/docs/screenshots/02-clown-evaluation-3.JPG" width="300" alt="OpenClown evaluation โ part 3">
### `/clown encore` โ improved response with feedback applied
<img src="https://raw.githubusercontent.com/Bubbletea98/openclown/main/docs/screenshots/03-clown-encore.JPG" width="300" alt="OpenClown encore โ improved response">
### Reply targeting
Reply to **any** AI response โ not just the latest โ and type `/clown` to evaluate that specific exchange:
```
[Reply to any OpenClaw message] /clown
โ ๐ Matched reply: "Yes! Toronto public pools are open on Sunday..."
โ Evaluates that specific response with full context
```
### Follow-up questions
OpenClown automatically includes up to 3 prior exchanges as context, so evaluators understand follow-up questions:
```
You: Which cities in Canada are best for startups?
OpenClaw: Vancouver and Toronto lead the pack...
You: How about Ottawa?
OpenClaw: Ottawa has a growing tech scene...
You: /clown
โ evaluators see both exchanges and understand
"How about Ottawa?" means "Is Ottawa good for startups?"
```
## Performers
OpenClown ships with 12 performers. Three are enabled by default:
| Performer | Emoji | Focus |
|-----------|-------|-------|
| **Philosopher** | ๐ญ | Assumptions, definitions, epistemic honesty |
| **Security Expert** | ๐ | Data exposure, API security, privacy |
| **Developer** | ๐ป | Implementation quality, error handling, efficiency |
Additional performers you can enable:
| Performer | Emoji | Focus |
|-----------|-------|-------|
| Ethicist | โ๏ธ | Fairness, inclusivity, potential harm |
| Fact Checker | ๐ | Accuracy, sources, hallucination detection |
| UX Designer | ๐จ | Information hierarchy, scannability, actionability |
| VC Investor | ๐ฐ | Value proposition, scalability, ROI |
| Comedian | ๐ | Absurdity, overthinking, unintentional humor |
| Shakespeare | ๐ญ | Narrative arc, emotional truth, prose quality |
| Conspiracy Theorist | ๐ฎ | Data provenance, hidden agendas, algorithmic bias |
| Grandparents | ๐ต๐ด | Practicality, common sense, well-being |
| Cat Expert | ๐ฑ | Efficiency, priorities, power dynamics |
## Managing your circus
Use `/clown circus` to see your current lineup. Enable or disable performers using their id or number:
```
/clown circus on comedian # Enable by id
/clown circus on 4,5 # Enable by number
/clown circus off philosopher # Disable by id
/clown circus off 1 # Disable by number
/clown circus toggle 1,3,8 # Switch multiple on/off in one command
/clown circus reset # Restore defaults (Philosopher, Security, Developer)
```
Your lineup is saved to `~/.openclaw/openclown/circus.json` and persists across restarts.
## Create your own performer
Don't see the perspective you need? Create your own through a guided conversational flow โ no code or config files needed:
```
You: /clown circus create A maritime law expert who evaluates
responses for legal accuracy around shipping regulations
OpenClaw: ๐ช Creating a new performer...
1. What specific aspects of maritime law should this
evaluator focus on?
2. Should the evaluation style be formal/checklist-based
or more conversational?
3. How severe should findings be โ advisory insights,
warnings, or critical errors?
...
You: /clown circus create Focus on UNCLOS compliance and
cargo liability. Formal style. Severity: warning.
OpenClaw: ๐ช Here's your performer draft:
โโโโโโโโโโโโโโโโโโโโโโ
โ๏ธ Maritime Law Expert [maritime]
Severity: โ ๏ธ Warning
Category: serious
โโโโโโโโโโโโโโโโโโโโโโ
/clown circus confirm โ save and enable
/clown circus preview โ see full definition
/clown circus create <changes> โ revise
/clown circus cancel โ discard
You: /clown circus confirm
OpenClaw: ๐ช New Performer Created!
โ๏ธ Maritime Law Expert [maritime]
โ
Saved and enabled.
```
Custom performers are saved to `~/.openclaw/openclown/skills/` and work exactly like built-in ones. You can also edit or delete them:
```
/clown circus edit maritime Make it focus more on cargo liability
/clown circus delete maritime
```
For the full command reference (configuration, manual SKILL.md format, all options), see the [Command Reference](https://github.com/Bubbletea98/openclown/blob/main/docs/commands.md).
## Multilingual support
OpenClown auto-detects the language of the user's request and evaluates in the same language. Supported locales: English, Chinese, Japanese, Korean, French, Spanish.
## Development
```bash
git clone https://github.com/Bubbletea98/openclown.git
cd openclown
npm install
npm run build
npm test
```
See [CONTRIBUTING.md](https://github.com/Bubbletea98/openclown/blob/main/CONTRIBUTING.md) for contribution guidelines.
## License
MIT
tools
Comments
Sign in to leave a comment