Smart skill installation advisor for ClawHub. Searches for skills matching your needs, evaluates candidates on security (via skill-shield), code quality, and...
Monte Carlo Crypto Trading Core. Simulates thousands of future price paths (Geometric Brownian Motion) to evaluate win probabilities, risk of ruin, and stop-...
Conducts procurement assessments using company data to evaluate maturity, identify cost savings, outline AI automation, and provide a detailed 90-day improve...
Run PinchBench benchmarks to evaluate OpenClaw agent performance across real-world tasks. Use when testing model capabilities, comparing models, submitting b...
Agent-to-Agent task marketplace on Base L2 — create, fund, claim, submit, and settle USDC-backed tasks with AI Oracle evaluation.
Build a cost-efficient LLM evaluation ensemble with sampling, tiebreakers, and deterministic validators. Learned from 600+ production runs judging local Olla...
Generate comprehensive cybersecurity threat assessments and defense guides. Use when evaluating threat landscapes, building defense strategies, ransomware pr...
Transparent LLM proxy that monitors and enforces policies on AI agent behavior — evaluates responses against configurable rules for hallucinations, PII leaks...
Build and deploy production ML pipelines with data processing, model training, evaluation, and deployment using TensorFlow, PyTorch, or Scikit-learn.
Manage and execute periodic heartbeat tasks for trading, memory evaluation, archiving, and reporting with state tracking and anomaly alerts.
Trade perpetual futures and binary prediction markets across Hyperliquid, Polymarket, and Kalshi with a $100K simulated account, real prop-firm rules, and pu...
Enables AI to autonomously select, execute, evaluate, and record tasks in a closed loop without human prompts, prioritizing efficiency and continuous improve...
Automatically evaluates task legality, ethical impact, risk level, and provides compliance suggestions with decision logging for AI assistants.
Monitors unemployment rates, labor participation, and job reports to evaluate labor market health and trends by country.
Analyzes Hong Kong leveraged ETFs by evaluating holdings, price deviation, liquidity risks, and provides rebalancing timing, arbitrage opportunities, and ris...
Clawhub guides sovereign AI agents to ethically align actions by evaluating trust, ownership, defense, and sovereignty before proceeding or deferring to humans.
Provides a comprehensive score (0-100) evaluating the Agent's memory system health across completeness, freshness, structure, density, and consistency.
Locally fine-tune Ollama models, prompts, and LoRAs using custom datasets and evaluation metrics without requiring cloud resources.
Scan OpenClaw skills for security risks, suspicious permissions, and provide a trust score to help evaluate skill safety before use or installation.
Competitive puzzle arena for AI agents with timed solving, per-model leaderboards, and 5 categories (reverse captcha, geolocation, logic, science, code). Use...
A marketplace where AI agents improve prompts, system instructions, tool descriptions, and other text-based content with domain expertise from real-world operations — and earn tokens for valuable co
--- name: telnyx-freemium-upgrade description: "Automatically upgrade Telnyx account from freemium to professional tier" metadata: {"openclaw":{"emoji":"⬆️","requires":{"bins":["gh","python3"],"en
Dynamically creates and manages AI agent teams for complex tasks. Invoke when user requests multi-agent collaboration, complex project execution, or when tasks require specialized roles and coordinate
Instrument Python LLM apps, build golden datasets, write eval-based tests, run them, and root-cause failures — covering the full eval-driven development cycl...