Bundle Pluginsource linked

olostepv1.0.0

Olostep for OpenClaw

olostep-web-agent·runtime olostep-web-agent·by @zeeshanadilbutt
openclaw bundles install clawhub:olostep-web-agent
Latest release: v1.0.0Download zip

Capabilities

Bundle format
clawhub
Host targets
openclaw
Runtime ID
olostep-web-agent
Security Scan
VirusTotalVirusTotal
Pending
View report →
OpenClawOpenClaw
Suspicious
medium confidence
Purpose & Capability
The declared capabilities (scrape, crawl, batch, answers, integrate) legitimately require network access and an Olostep API key and may need access to a user's repo to integrate SDKs — the capabilities themselves are coherent with the described features. However, the package metadata lists no required env vars or config paths even though the SKILL.md and mcp.json show OLOSTEP_API_KEY and instructions to add an MCP server to OpenClaw; that's an inconsistency.
!
Instruction Scope
SKILL.md (especially skills/integrate/SKILL.md) instructs the agent to: inspect the user's project files (package.json, pyproject.toml, requirements.txt, .env), modify or create a .env, write integration code, and (in the top-level docs) add an MCP server entry to OpenClaw config. Those are legitimate for an integration flow but touch arbitrary project files and agent config — and the metadata does not declare this file access. The instructions also direct network calls to api.olostep.com and to install packages (npm/pip) which will execute external code if the user follows them.
Install Mechanism
There is no automated install spec bundled with the registry entry (instruction-only). That lowers installer risk because nothing arbitrary is automatically downloaded by the skill itself. However, SKILL.md recommends installing third-party packages (npm/pip) and adding an MCP server that will run 'npx olostep-mcp' — those are external installs the user would perform, so review of the referenced upstream packages (npm/pypi) is recommended.
!
Credentials
The skill repeatedly references OLOSTEP_API_KEY (mcp.json, SKILL.md examples, SDK usage, and CI workflows) and the integrate skill instructs storing an API key in the project's .env. Yet the registry metadata claims 'Required env vars: none' and 'Primary credential: none'. This mismatch is the primary red flag: the skill will be nonfunctional without that secret, and the metadata fails to disclose it. Tests and CI also reference a secret (GitHub workflow uses secrets.OLOSTEP_API_KEY), showing that live API credentials are expected for full functionality.
Persistence & Privilege
always:false and model invocation is enabled (normal). The skill asks users to add an MCP server entry to their OpenClaw configuration and to store an API key in a .env file — these are persistent changes the user must approve explicitly. The skill does not request forced always-on or elevated platform-wide privileges in the registry metadata, but the SKILL.md's recommended config edits mean the user should be deliberate when applying them.
What to consider before installing
This package advertises a web-scraping and integration toolkit that expects an Olostep API key and instructs the agent to read and modify project files and your OpenClaw config — but the registry metadata does not declare that key or those config paths. Before installing or enabling the skill: (1) verify the upstream packages (npm 'olostep', 'olostep-mcp' and Python 'olostep') on their official registries and review their code and privacy/terms; (2) be cautious about giving the agent permission to read/write your repo (.env, package.json, source files) — run integration steps manually in an isolated test repo first; (3) if you must provide an API key, prefer a scoped/test key and avoid using high-privilege production credentials; (4) consider running the MCP server in an isolated environment (separate process/container) and review network endpoints contacted (api.olostep.com); and (5) ask the publisher to correct the package metadata to declare OLOSTEP_API_KEY and any required config paths so the security implications are explicit.

Verification

Tier
source linked
Scope
artifact only
Summary
Validated package structure and linked the release to source metadata.
Commit
78e4673a5095
Tag
main
Provenance
No
Scan status
pending

Tags

latest
1.0.0

Olostep for OpenClaw

OpenClaw Plugin Version License Skills MCP Tools CI

Give your OpenClaw agent the live web. Scrape any URL to clean markdown, search Google with structured results, crawl entire sites, batch-process up to 10,000 URLs in parallel, and get AI-synthesized answers with citations — all from inside your agent workflow.

This plugin provides 13 skills for high-level agent workflows and a 9-tool MCP server for direct tool access. Handles JS rendering, anti-bot bypass, CAPTCHAs, and residential proxies automatically.


Installation

Install from ClawHub:

clawhub install olostep

Or add the MCP server directly to your OpenClaw configuration:

{
  "mcpServers": {
    "olostep": {
      "command": "npx",
      "args": ["-y", "olostep-mcp"],
      "env": {
        "OLOSTEP_API_KEY": "your-api-key-here"
      }
    }
  }
}

Get your free API key at olostep.com/auth — 500 requests/month, no credit card.


Skills

Core Data Skills

SkillWhat it does
scrapeExtract clean markdown, HTML, JSON, or text from any URL. Full browser rendering with anti-bot bypass. Supports geo-targeting, browser actions (click, scroll, type), and pre-built parsers.
searchThree search tools in one: answers for AI-synthesized results, google_search for raw SERP data, and get_website_urls for finding pages within a specific domain.
crawlFollow links from a starting URL and scrape every page discovered. Set max_pages, include/exclude URL patterns, and control crawl depth.
batchScrape up to 10,000 URLs in parallel. All pages rendered concurrently. Use custom_id to map results back to sources.
mapDiscover every URL on a website instantly. Filter by glob patterns, rank by search query relevance, limit with top_n.
answersAsk a question, get an AI-synthesized answer from live web sources with citations. Pass a json parameter to get structured output in any shape you define.

Workflow Skills

SkillWhat it does
researchMulti-source competitive intelligence and tool comparisons. Combines search, scrape, and batch to deliver structured, cited analysis with a recommendation.
debug-errorPaste a stack trace. The agent searches the live web for that exact error, scrapes the GitHub issue or StackOverflow thread, and translates it into a fix for your codebase.
docs-to-codeScrape real API documentation and write working integration code from what is actually there — not from stale training data.
migrate-codeScrape the migration guide and changelog, extract breaking changes, and refactor your code based on the real documentation.
extract-schemaTurn any unstructured webpage into typed JSON matching a TypeScript interface, JSON schema, or database model.
integrateAuto-detect your stack (language, framework, AI toolkit) and write a complete Olostep SDK integration — install, client setup, tool wiring, and verification.
setupConfigure the Olostep API key and verify the connection. Troubleshooting for common issues.

MCP Tools

The bundled MCP server (olostep-mcp) exposes 9 tools callable by any MCP-compatible agent:

ToolDescription
scrape_websiteScrape a single URL — markdown, HTML, JSON, or text output
get_webpage_contentRetrieve a webpage as clean markdown
search_webSearch the web, get AI-synthesized answers
google_searchStructured Google SERP data (organic results, PAA, knowledge graph)
answersAI-powered answers with citations and optional structured JSON
batch_scrape_urlsBatch scrape up to 10,000 URLs in parallel
create_crawlCrawl a website by following links from a start URL
create_mapDiscover all URLs on a website with filtering
get_website_urlsSearch and retrieve relevant URLs from a specific domain

Real Developer Workflows

Debug an obscure error

Your agent searches for the exact error message across GitHub issues and documentation, scrapes the most relevant threads, and delivers a fix — not a generic suggestion, an actual fix based on what people who hit the same error actually did.

Write code from live docs

AI models hallucinate API parameters and use deprecated methods. The docs-to-code skill scrapes the current documentation and writes code from what is really there. No stale training data, no guessing.

Research before choosing a tool

Comparing ORMs? Evaluating payment providers? The research skill searches multiple sources, scrapes actual pricing and feature pages, and returns a structured comparison with a recommendation and citations.

Ingest an entire docs site for RAG

map → discover all URLs on the docs site
batch → scrape every page in parallel as clean markdown
→ feed into your vector store

Extract structured data at scale

map → find all product/listing URLs
batch → scrape with a pre-built parser for typed JSON
→ pipe into your database, API, or seed files

Migrate to a new framework version

The migrate-code skill scrapes the official migration guide, extracts every breaking change with before/after patterns, and refactors your code based on the real documentation.


SDK Quick Reference

Node.js / TypeScript

npm install olostep
import Olostep from 'olostep';

const client = new Olostep({ apiKey: process.env.OLOSTEP_API_KEY });

// Scrape a page to markdown
const scrape = await client.scrapes.create('https://example.com');
console.log(scrape.markdown_content);

// Batch scrape with custom IDs
const batch = await client.batches.create([
  { url: 'https://example.com/page-1', customId: 'p1' },
  { url: 'https://example.com/page-2', customId: 'p2' },
]);

// Crawl a site
const crawl = await client.crawls.create({
  url: 'https://docs.example.com',
  maxPages: 50,
  maxDepth: 3,
});

// Discover URLs
const map = await client.maps.create({
  url: 'https://example.com',
  topN: 100,
  searchQuery: 'pricing',
});

Python

pip install olostep
from olostep import Olostep

client = Olostep()  # reads OLOSTEP_API_KEY from env

# Scrape
result = client.scrapes.create(url_to_scrape="https://example.com")
print(result.markdown_content)

# Batch
batch = client.batches.create(urls=["https://example.com", "https://example.org"])

# Crawl
crawl = client.crawls.create(start_url="https://docs.example.com", max_pages=50)

# Map
site_map = client.maps.create(url="https://example.com")

# AI Answers with structured output
answer = client.answers.create(
    task="Compare Stripe vs Square pricing",
    json_format={"providers": [{"name": "", "pricing": "", "best_for": ""}]},
)

REST API

Base URL: https://api.olostep.com/v1 Auth: Authorization: Bearer <API_KEY>

MethodEndpointPurpose
POST/v1/scrapesScrape a single URL
GET/v1/scrapes/:idGet scrape result
POST/v1/batchesStart a batch (up to 10k URLs)
GET/v1/batches/:idGet batch status
GET/v1/batches/:id/itemsGet batch results
POST/v1/crawlsStart a crawl
GET/v1/crawls/:idGet crawl status
GET/v1/crawls/:id/pagesGet crawled pages
POST/v1/mapsMap a website's URLs
POST/v1/answersGet AI-powered answers
GET/v1/answers/:idGet answer result
GET/v1/retrieveRetrieve content by ID

Pre-built Parsers

Use with parser to get structured JSON from specific site types:

ParserUse case
@olostep/google-searchGoogle SERP (organic results, knowledge graph, PAA)
@olostep/amazon-it-productAmazon product data (price, rating, features)
@olostep/extract-emailsExtract email addresses from any page
@olostep/extract-calendarsExtract calendar events
@olostep/extract-socialsExtract social media profile links

Framework Integrations

This plugin works with OpenClaw directly. If you are building outside of OpenClaw, Olostep has dedicated SDKs and integrations:

FrameworkPackageWhat you get
LangChain (Python)langchain-olostepscrape_website, answer_question, scrape_batch, crawl_website, map_website
LangChain (TypeScript)langchain-olostepOlostepScrape, OlostepAnswer, OlostepBatch, OlostepCrawl, OlostepMap
CrewAIcrewai-olostepolostep_scrape_tool, olostep_answer_tool, olostep_batch_tool, olostep_crawl_tool, olostep_map_tool
Mastra@olostep/mastraFull Olostep toolset for Mastra agents
MCP (Cursor, Claude Desktop)olostep-mcp9 MCP tools for any MCP-compatible client
Vercel AI SDKolostepCreate AI tools that wrap scrape and search
OpenAI function callingolostepDefine web_search and scrape_url functions

Testing

The plugin ships with two test suites:

Static integrity tests (run on every push and PR):

pip install pytest
pytest tests/test_skill_integrity.py -v

Validates: frontmatter fields, MCP tool references, Python code block syntax, and absence of deprecated parameters across all 13 skill definitions.

Live API smoke tests (run on push to main):

OLOSTEP_API_KEY=your_key pytest tests/test_live_api.py -v --timeout=60

Verifies scrape, map, answers, and batch workflows against the live Olostep API.


Project Structure

olostep-clawhub-plugin/
  .clawhub/
    plugin.json          # Plugin metadata, keywords, logo reference
  .github/
    workflows/
      integrity.yml      # CI: static checks on PR, live tests on push to main
  skills/
    answers/SKILL.md     # AI-powered answers with citations
    batch/SKILL.md       # Batch scrape up to 10k URLs
    crawl/SKILL.md       # Autonomous website crawling
    debug-error/SKILL.md # Error debugging from live web
    docs-to-code/SKILL.md # Write code from scraped docs
    extract-schema/SKILL.md # Structured data extraction
    integrate/SKILL.md   # Auto-detect stack, write SDK integration
    map/SKILL.md         # URL discovery and site mapping
    migrate-code/SKILL.md # Framework migration from docs
    research/SKILL.md    # Multi-source competitive research
    scrape/SKILL.md      # Single-page scraping
    search/SKILL.md      # Web search (3 modes)
    setup/SKILL.md       # API key configuration
  tests/
    conftest.py          # Shared fixtures, known MCP tools, stale params
    test_skill_integrity.py  # Static validation of all SKILL.md files
    test_live_api.py     # Live API smoke tests
  mcp.json               # MCP server configuration
  LICENSE                 # MIT
  README.md

Links

License

MIT — see LICENSE for details.