Integration
Plugin Crw
CRW web scraping plugin for OpenClaw — scrape, crawl, and map websites
Install
openclaw plugins install openclaw-plugin-crw
Configuration Example
{
"plugins": {
"crw": {
"apiKey": "crw_live_..."
}
}
}
README
# openclaw-plugin-crw
[](https://www.npmjs.com/package/openclaw-plugin-crw)
[](https://opensource.org/licenses/MIT)
[CRW](https://github.com/us/crw) web scraping plugin for [OpenClaw](https://github.com/openclaw/openclaw). Gives your AI agents the ability to scrape, crawl, and map websites.
## Installation
```bash
openclaw plugins install openclaw-plugin-crw
```
## Setup — Pick One
### Option A: Cloud ([fastcrw.com](https://fastcrw.com)) — Quickest Start
No server to install. [Sign up at fastcrw.com](https://fastcrw.com) and get **500 free credits** to start scraping:
```json
{
"plugins": {
"crw": {
"apiKey": "crw_live_..."
}
}
}
```
That's it — cloud is the default. Your agents can now scrape, crawl, and map any website.
### Option B: Self-hosted with binary (free, no limits)
Single binary, ~15 MB download, ~6 MB idle RAM. No Docker needed.
```bash
curl -fsSL https://raw.githubusercontent.com/us/crw/main/install.sh | bash
crw # starts on http://localhost:3000
```
```json
{
"plugins": {
"crw": {
"apiUrl": "http://localhost:3000"
}
}
}
```
### Option C: Self-hosted with Docker
```bash
docker run -d -p 3000:3000 ghcr.io/us/crw:latest
```
```json
{
"plugins": {
"crw": {
"apiUrl": "http://localhost:3000"
}
}
}
```
## Tools
| Tool | Description |
|------|-------------|
| `crw_scrape` | Scrape a single URL and get clean markdown |
| `crw_crawl` | BFS crawl a website, collect content from multiple pages |
| `crw_map` | Discover all URLs on a website via sitemap + link traversal |
## How It Works
Once installed and configured, your OpenClaw agents use the CRW tools automatically:
### Scrape a page
**User (via WhatsApp/Telegram/Discord):** "Summarize this article: https://example.com/blog/post"
**Agent uses `crw_scrape`** → gets clean markdown → summarizes → responds.
### Crawl an entire site
**User:** "Research everything on docs.example.com"
**Agent uses `crw_crawl`** → discovers and scrapes all pages → synthesizes findings → responds.
### Discover site structure
**User:** "What pages does example.com have?"
**Agent uses `crw_map`** → returns all discovered URLs via sitemap + link traversal.
## Example: URL Scraper Bot
A simple OpenClaw agent that scrapes any URL users send:
1. User sends a URL via WhatsApp/Telegram
2. Agent detects the URL and calls `crw_scrape`
3. CRW fetches the page (using [fastcrw.com](https://fastcrw.com) cloud or your local instance)
4. Agent receives clean markdown — no HTML noise, no nav/footer
5. Agent summarizes and replies
With fastcrw.com, no infrastructure needed — just plug in your API key and go.
## Compared to Firecrawl Plugin
| Feature | CRW Plugin | Firecrawl Plugin |
|---------|-----------|-----------------|
| Cloud option | [fastcrw.com](https://fastcrw.com) | firecrawl.dev |
| Self-hosted | Yes (single binary, ~6 MB) | Complex (5+ containers) |
| API key required | No (self-hosted) | Yes (always) |
| Idle RAM | ~6 MB | ~500 MB+ |
| Avg latency | 833ms | 4,600ms |
| Cost (self-hosted) | $0 | $0 but heavy infra |
## License
MIT
integration
Comments
Sign in to leave a comment