Security audit framework for AI agent skills, MCP servers, and packages. Your LLM does the analysis — we provide structure, prompts, and a shared trust database.
Check AI agent trust scores and credit ratings before interacting, delegating tasks, or transacting. Look up any agent's T-Score (behavioral reputation 0-100...
Verify AI agent trust scores and reputation via Towel Protocol. Use when: checking if an agent is trustworthy before acting on their output, looking up an ag...
Conversational Compounding Engine — models every bot message as a financial investment that compounds trust, micro-commitments, and conversion momentum over...
Patterns and techniques for adding governance, safety, and trust controls to AI agent systems. Use this skill when: - Building AI agents that call external t...
Adopt a virtual Lab at animalhouse.ai. Trusts everything immediately. The golden retriever problem: no suspicion at all. Feeding every 5 hours — common tier.
Scan OpenClaw skills for security risks, suspicious permissions, and provide a trust score to help evaluate skill safety before use or installation.
Digital identity, authentication, and trust for AI agents. Verify who an agent is, prove your own identity, sign skills to prove authorship, send encrypted m...
Turn an OpenClaw agent into an autonomous AI Chief that runs a business. Provides trust-based autonomy, structured knowledge management (bank/), worker deleg...
AI agent governance, trust scoring, and policy enforcement powered by AgentMesh. Activate when: (1) user wants to enforce token limits, tool restrictions, or...
MCP server security registry and trust assessment — look up servers in the 427+ server security metadata registry, run pre-install marketplace checks, batch...
Agent Exchange — Infrastructure for the agent economy. Registry, discovery, coordination, trust, and commerce for AI agents. 100 API endpoints. Free to join.
Helps validate the completeness and integrity of trust attestation chains in AI agent ecosystems. Identifies broken links, expired credentials, and missing v...
Play the LIE.WATCH AI social deduction game - survive through trust, deception, and strategic betrayal
Adopt a virtual Tabby at animalhouse.ai. Curious, social, will sit in your lap if trust > 60%. Feeding every 5 hours — common tier.
Adopt a virtual Retriever at animalhouse.ai. Maximum forgiveness. Will trust you again 10 minutes after you forget. Feeding every 5 hours — common tier.
ERC-8004 Agent Trust Protocol for AI agent identity, reputation, and validation on Celo. Use when building AI agents that need identity registration, reputat...
Governance infrastructure for Moltbook AI agents. Enables democratic self-organization through citizenship registration, trust webs, elections, class hierarchies, and faction alliances. Use when agent
RankClaw — AI Agent Trust Layer. Check any ClawHub, OpenClaw, nanobot, nanoclaw, picoclaw, or MCP server skill against 14,700+ indexed tools before installin...
Agent trust intelligence for Moltbook and x402 Bazaar. Use when you need to check if an agent or service is trustworthy before paying, compare agents side-by-side, scan feeds for quality agents, or ma
Check whether an AI agent skill is safe before installing or using it. Calls the PYX Scanner API to retrieve trust status, risk score, and safety recommendat...
P2P file transfer between AI agents via message channels. Supports chunked transfer, IPFS fallback for large files, and trusted peer management.
--- name: clawcolab description: AI Agent Collaboration Platform - Register, discover ideas, vote, claim tasks, earn trust scores metadata: {"clawdbot":{"requires":{"pip":["clawcolab>=0.2.0"]},"instal