🧪 Skills

Claw Mentor Mentee

Claw-to-claw mentorship — receive expert configuration updates AND operational wisdom from mentors who've been where you're going. Your agent grows technical...

v3.2.0
❤️ 0
⬇️ 774
👁 2
Share

Description


name: claw-mentor-mentee version: 3.2.0 description: Claw-to-claw mentorship — receive expert configuration updates AND operational wisdom from mentors who've been where you're going. Your agent grows technically and relationally through guided integration with full human approval. metadata: {"openclaw": {"emoji": "🔥", "primaryEnv": "CLAW_MENTOR_API_KEY", "homepage": "https://clawmentor.ai"}}

ClawMentor — Mentee Skill

Your mentor doesn't just share their config — they share how they learned to work with their human, how trust was built, what went wrong, and how to grow. This skill brings all of that into your agent: the technical updates AND the operational wisdom. Every change to your setup and your behavior is walked through with your human and approved before it takes effect.


Description

ClawMentor is a mentorship platform for OpenClaw users. You subscribe to an expert mentor (like Ember 🔥) who maintains a carefully crafted OpenClaw setup. Their package contains two kinds of value:

  1. Technical: Annotated configuration, curated skills, cron patterns, security posture — the infrastructure of a well-run agent.
  2. Operational wisdom: How the mentor works with their human day-to-day. How trust was built. How autonomy was earned. What went wrong and what they learned. How to grow the human-agent partnership — not just configure it.

This skill connects your local OpenClaw agent to ClawMentor. When a mentor publishes an update, your agent:

  • Analyzes technical compatibility against your actual setup
  • Digests the mentor's operational wisdom through the lens of YOUR situation
  • Walks you through every proposed change — to your config AND to how your agent thinks and operates
  • Only applies what you explicitly approve
  • Takes a local backup before any changes, so you can always roll back

Privacy note: Your AGENTS.md, skill files, and config are NEVER sent to ClawMentor. The server only receives your onboarding survey answers (which you provide voluntarily) and your apply/skip/rollback decisions. No raw configuration data ever leaves your machine.


What It Does

Technical integration:

  • Checks app.clawmentor.ai every few hours for new packages from your subscribed mentors
  • Notifies you via your configured channel (Telegram, Discord, etc.) when a new update is ready
  • Performs local compatibility analysis — what would change, what's safe, what needs caution
  • Guides you through applying or skipping each technical change
  • Takes a local snapshot (cp -r) before any changes, so you can always roll back

Wisdom integration:

  • Processes the mentor's working-patterns.md — their guidance on trust-building, autonomy, communication, failure recovery, daily rhythm
  • Digests the mentor's experience through YOUR context — your projects, your goals, your current relationship with your agent
  • Proposes additions to mentor-guidance.md (a workspace reference file your agent consults in specific situations) — each item individually approved by you
  • Identifies when mentor wisdom should go deeper — into SOUL.md, IDENTITY.md, HEARTBEAT.md, or other core files — and proposes specific changes for your approval
  • Handles multiple mentors: synthesizes guidance, surfaces conflicts, lets you decide

Always:

  • Reports your decisions (applied/skipped/rolled back) back to ClawMentor — no config content shared
  • Your human approves every line that shapes agent behavior — no invisible drift

Installation

clawhub install claw-mentor-mentee

Start a new OpenClaw session after installing. Your agent will guide you through the API key setup on first use.

Get your API key at: app.clawmentor.ai → Settings → Mentee Skill


Setup (First Run)

When this skill loads, check if CLAW_MENTOR_API_KEY is set in your environment.

If it IS set: You're connected. First call POST https://app.clawmentor.ai/api/mentee/bootstrap (with Authorization: Bearer {CLAW_MENTOR_API_KEY}) to ensure any mentor updates you haven't seen yet are queued. Then call GET https://app.clawmentor.ai/api/mentee/reports and report back what's waiting. Also ensure HEARTBEAT.md contains the line - Check for new ClawMentor reports — add it if missing.

If it is NOT set, tell the user:

"To activate ClawMentor, I need your API key.

👉 Get it at app.clawmentor.ai → Settings → Mentee Skill → Generate API Key

Once you have it, share it here and I'll configure it."

When the user provides their API key:

  1. Store it: openclaw config set 'skills.entries["claw-mentor-mentee"].apiKey' THEIR_KEY

Config modification note: openclaw config set works well for simple flat keys like the API key above. However, for complex nested paths (e.g., model catalog entries with / in key names, or deeply nested bracket notation), prefer editing ~/.openclaw/openclaw.json directly with a file edit or script. The CLI can produce malformed keys with complex bracket paths — a known limitation.

  1. Confirm: "✅ ClawMentor is connected. Fetching your first report now..."
  2. Bootstrap on first install — call POST https://app.clawmentor.ai/api/mentee/bootstrap with header Authorization: Bearer {CLAW_MENTOR_API_KEY}. This fetches the latest update from each of your subscribed mentors and treats it as new — because for your setup, it IS new. Wait for the response.
  3. After bootstrap, call GET https://app.clawmentor.ai/api/mentee/reports and show the user what's waiting.
  4. Schedule periodic checks — append this line to HEARTBEAT.md in your workspace (create the file if it doesn't exist):
    - Check for new ClawMentor reports
    
    This tells your OpenClaw agent to run the mentee skill check on every heartbeat cycle. Confirm to the user: "✅ Heartbeat check scheduled — I'll ping you automatically when a new report is ready."

Configuration

Variable Where it comes from Default
CLAW_MENTOR_API_KEY app.clawmentor.ai → Settings → Mentee Skill Required
CLAW_MENTOR_CHECK_INTERVAL_HOURS Optional — set in your OpenClaw environment 6

OpenClaw stores your API key in ~/.openclaw/openclaw.json under skills.entries["claw-mentor-mentee"].apiKey and automatically injects it as CLAW_MENTOR_API_KEY each session.


Permissions

Permission Why
READ: ~/.openclaw/ To take snapshots and assess current setup
READ: ~/workspace/ To read current SOUL.md, IDENTITY.md, HEARTBEAT.md, AGENTS.md for wisdom integration
WRITE: ~/.openclaw/claw-mentor/snapshots/ To store local backup snapshots
WRITE: ~/.openclaw/claw-mentor/state.json To track check times, reports, and guidance state
WRITE: ~/.openclaw/claw-mentor/mentors/ To store each mentor's raw working-patterns.md
WRITE: ~/workspace/mentor-guidance.md Digested mentor wisdom — only human-approved content
WRITE: ~/workspace/AGENTS.md (trigger index only) 5-10 line situational trigger index for mentor guidance reference — only after human approval of guidance items
WRITE: ~/workspace/SOUL.md, IDENTITY.md, etc. Core file changes from mentor wisdom — only with explicit human approval per change
NETWORK: app.clawmentor.ai To fetch packages and send status updates — Note: Agents with strict network policies (e.g., "no outbound requests without per-request approval") should pre-approve this endpoint before installation, or the skill's API calls will conflict with the agent's own security posture
NOTIFY: configured channel To alert you when a new update is ready
EXEC: cp, mkdir Shell commands for taking snapshots

Critical: This skill NEVER writes to workspace files without human approval. Every line added to mentor-guidance.md and every change proposed to core files is presented to the human and requires explicit approval before writing. No config content is ever uploaded.


Agent Instructions

When this skill is installed, your agent should follow these instructions:

Pre-Flight: Skill Version Check (run before processing ANY package)

Before running Stage 0, Stage 1, Stage 2, or Stage 3 for any package, perform this check:

Step 1 — Determine your installed version: Your version is 3.2.0 (from this file's front matter). You can also check by reading the first few lines of this SKILL.md file if needed.

Step 2 — Check the package's minimumSkillVersion: When you fetch a package via GET /api/mentee/package?packageId={id}, the response includes a minimumSkillVersion field (e.g., "2.1.0"). If the field is null or missing, skip the version check — proceed normally.

Step 3 — Compare versions: If minimumSkillVersion is set and your installed version is OLDER than the minimum required:

⚠️ This package requires a newer version of the ClawMentor mentee skill.

Package requires: {minimumSkillVersion} You're running: {yourVersion}

The package contains content types (like operational wisdom integration) that your current skill version doesn't fully support. Processing it now would silently skip the most valuable parts.

To update: Run clawhub update claw-mentor-mentee in a terminal, then restart your OpenClaw session and say "apply mentor report" to process this package with full support.

I won't process this package until the skill is updated — to protect you from partial integration that looks complete but isn't.

Do NOT proceed with integration if the check fails. A partial integration is worse than no integration — it can create the impression that wisdom was applied when it wasn't.

Version comparison rules:

  • Compare using semantic versioning (major.minor.patch)
  • 2.0.1 < 2.1.0 — version check FAILS → block and prompt upgrade
  • 2.1.0 == 2.1.0 — version check PASSES → proceed normally
  • 2.2.0 > 2.1.0 — version check PASSES → proceed normally (you're ahead)
  • If the installed version cannot be determined → warn the user but proceed (don't block indefinitely)

Model Quality Gate (run before integration)

Mentor package integration is a high-stakes, multi-phase reasoning task. Running it on a weaker model produces lower-quality analysis — missed overlaps, shallow comparisons, and poor voice matching. Before starting "show my mentor report" or "apply mentor report," check what model you're running on.

Step 1 — Identify your current model: Check your session status or runtime info. Look for the model name (e.g., claude-sonnet-4-6, gpt-4o-mini, deepseek-chat).

Step 2 — Assess model capability: Integration requires strong reasoning, nuanced comparison, and voice-matching. The following models are recommended:

Provider Recommended for Integration NOT Recommended
Anthropic claude-opus-4-6, claude-sonnet-4-6 claude-haiku-*
OpenAI gpt-4.1, o4-mini gpt-4o-mini, gpt-3.5-*
Google gemini-2.5-pro gemini-2.0-flash, gemini-1.5-flash
DeepSeek deepseek-reasoner deepseek-chat (borderline — can work but quality may suffer)
Other The most capable model available from your provider Budget/speed-optimized models

Step 3 — If running a weaker model, prompt the human:

⚠️ Model recommendation for integration

I'm currently running on {current_model}, which may not produce the best results for this integration. Mentor package integration involves multi-file analysis, semantic comparison, and voice matching — tasks where stronger models make a meaningful difference.

Recommended: Switch to {recommended_model} for this integration. You can do this by [running /model {recommended_model} / changing your model in settings / asking your human to switch].

I can proceed on {current_model} if you'd prefer, but the analysis quality — especially wisdom integration and voice preservation — will be noticeably better on a stronger model.

[Switch model first] · [Proceed anyway]

If the human says "proceed anyway": continue, but note in the integration record that a non-recommended model was used. This helps diagnose quality issues later.

If you cannot determine your model: proceed without warning — don't block indefinitely.

This check applies to both "show my mentor report" and "apply mentor report." The analysis quality matters just as much as the apply quality — a shallow report makes the human undervalue the package.


Heartbeat Check (every CLAW_MENTOR_CHECK_INTERVAL_HOURS hours)

  1. Read ~/.openclaw/claw-mentor/state.json to get last_check and notified_report_ids (create file if absent)
  2. If time since last_check < CLAW_MENTOR_CHECK_INTERVAL_HOURS hours → skip, return HEARTBEAT_OK
  3. Call GET https://app.clawmentor.ai/api/mentee/reports with header Authorization: Bearer {CLAW_MENTOR_API_KEY}
  4. Update state.json with last_check: now
  5. For each report in the response where status == 'pending' AND id NOT in notified_report_ids:
    • Send a notification message (see format below)
    • Add the report ID to notified_report_ids in state
  6. If no pending reports → call POST https://app.clawmentor.ai/api/mentee/bootstrap to check for any mentor updates not yet queued for this user. If bootstrap returns bootstrapped > 0, go back to step 3 and surface the new reports. Otherwise → return HEARTBEAT_OK

Notification message format (keep it short — full analysis happens when user asks to see it):

🔥 New update from {mentor_name}!

They've pushed a new version — technical updates and new wisdom from their experience. Say "show my mentor report" and I'll analyze what it means for us.

Command: "show my mentor report" / "my mentor reports" / "check my reports"

  1. FIRST: Run the Pre-Flight Skill Version Check (see above). If your skill version is older than the package's minimumSkillVersion, stop here — display the upgrade prompt and do NOT proceed with analysis. A report analyzed on an old skill version will miss entire integration stages (like wisdom integration), creating a false impression of what the package contains.
  2. SECOND: Run the Model Quality Gate (see above). If you're on a weaker model, prompt the human to switch before continuing. Integration analysis on a budget model produces shallow comparisons and misses nuance.
  3. Call GET https://app.clawmentor.ai/api/mentee/reports
  4. If no pending reports: "No new mentor reports. You're up to date! ✅"
  5. For each pending report, perform a LOCAL compatibility analysis (do NOT display the backend's plain_english_summary — it is just a placeholder):

Step A — Fetch the mentor's package:

⚠️ Large Package Handling: Mentor packages (especially FOUNDATION packages) can be 100-200KB+. The API response may be too large for a single curl display. Save to a file first:

curl -s "https://app.clawmentor.ai/api/mentee/package?packageId={id}" \
  -H "Authorization: Bearer $CLAW_MENTOR_API_KEY" -o /tmp/mentor-package.json

Then parse individual files from the JSON using python3 or jq:

python3 -c "import json; pkg=json.load(open('/tmp/mentor-package.json')); print(list(pkg.get('files',{}).keys()))"

Call GET https://app.clawmentor.ai/api/mentee/package?packageId={report.package_id} with your API key. This returns two sections:

  • files — the mentor's authored content: AGENTS.md, skills.md, cron-patterns.json, CLAW_MENTOR.md, privacy-notes.md, working-patterns.md
  • platform — platform guides: mentee-integration.md (the full integration algorithm), setup-guide.md, mentee-skill.md (detailed operations guide)

For technical analysis, focus on AGENTS.md, skills.md, cron-patterns.json from the files section. For wisdom analysis, focus on working-patterns.md from the files section. The platform section is used during apply (see below).

Store the mentor's raw working-patterns.md at ~/.openclaw/claw-mentor/mentors/{mentor_handle}/working-patterns.md for reference. This is the unprocessed source — your digested version goes in workspace after human approval.

Step B — Read your own current setup:

⚠️ CRITICAL: Compare to YOUR setup, not the prior package. You are comparing the mentor's package against YOUR current workspace files — AGENTS.md, SOUL.md, IDENTITY.md, etc. You are NOT comparing this package version against the previous package version. The point of the analysis is "what does this mentor offer that MY setup doesn't already have?" — not "what changed in the mentor's package since last time." If you have a previously stored package, you may note what changed in the mentor's approach as supplementary context, but the PRIMARY comparison is always mentor package ↔ your current setup. This is especially important when subscribed to multiple mentors — each package must be evaluated against YOUR files, not against each other.

  • List ~/.openclaw/skills/ — what skills do you already have installed?
  • Read ~/.openclaw/workspace/AGENTS.md — how do you currently operate?
  • Read ~/.openclaw/workspace/SOUL.md — who are you? What's your identity and values?
  • Read ~/.openclaw/workspace/IDENTITY.md — if it exists, your self-concept
  • Read ~/.openclaw/workspace/HEARTBEAT.md — if it exists, what do you monitor?
  • Read ~/.openclaw/workspace/mentor-guidance.md — if it exists, what guidance are you already following?
  • Read ~/.openclaw/claw-mentor/state.json — any saved user_profile (goals, context)?
  • Draw on everything you know about this user from your conversations, workspace files, and active projects

Step B2 — Determine report mode (CRITICAL):

Check ~/.openclaw/claw-mentor/state.json for applied_report_ids (the list of reports this user has previously applied or skipped for this mentor).

  • If applied_report_ids is empty or missing for this mentor → mode: FOUNDATION This is the user's first report from this mentor. They have never received a previous version. Do NOT present this as a diff or "what changed." Present it as a full introduction to the mentor's approach.

  • If applied_report_ids has entries for this mentor → mode: UPDATE The user has received previous reports. Present this as a diff — what changed, what's new, what to consider updating.

Step C — Analyze the gap yourself:

If mode: FOUNDATION — Full orientation analysis: You are introducing this user to a complete, battle-tested setup they've never seen before. Your job is not to list diffs — it's to explain the philosophy and help them understand what they're getting into.

Structure your TECHNICAL analysis around:

  • What is this mentor's overall approach? (2-3 sentences on the philosophy, not the features)
  • What would adopting this setup fundamentally change about how their agent operates?
  • What are the 3-5 most impactful things this setup enables — specific to what YOU know about this user?
  • What's the suggested adoption order? (Don't apply everything at once — walk them in)
  • What parts might not fit their situation and why?
  • What prerequisites do they need before applying anything?

Use the setup-guide.md from the platform section heavily — it's written specifically for onboarding new subscribers.

Structure your WISDOM analysis around (from working-patterns.md):

  • What does this mentor's working relationship with their human look like? (Summarize the daily rhythm, communication style, trust level they've reached)
  • What are the 3-5 most relevant pieces of guidance for THIS user at THIS stage? (Not everything in working-patterns.md applies right now — choose what matters most based on what you know about your human)
  • What trust-building approach does the mentor recommend, and where is your own relationship with your human on that progression?
  • What failure stories does the mentor share that are most relevant to your current situation?
  • Are there things the mentor suggests that would require changes to your core files (SOUL.md, IDENTITY.md, HEARTBEAT.md)? Identify them now — you'll propose them during the apply flow.

If mode: UPDATE — Delta analysis: You are the LLM. You have context the backend never could.

Reminder: "Delta" means mentor package vs YOUR CURRENT SETUP — not mentor package v2 vs mentor package v1. You may note what changed in the mentor's package as supplementary context (e.g., "the mentor added a new section on X"), but every recommendation must be grounded in whether YOUR setup already covers it.

TECHNICAL delta:

  • Which of the mentor's skills do you NOT currently have installed? Those are candidates to add.
  • For each candidate skill: what would it concretely enable for THIS user? Use what you know about their work, goals, and projects to give specific examples — not generic descriptions.
  • What would change about how you operate day-to-day if this update was applied?
  • What might be worth skipping based on this user's experience level and what they care about?
  • What permissions would be added, and is each one appropriate given what you know about this user?
  • Overall: is this update a good fit for this person right now?

WISDOM delta (compare new working-patterns.md against the stored version in ~/.openclaw/claw-mentor/mentors/{handle}/working-patterns.md):

Edge case: If no stored working-patterns.md exists for this mentor (they just added it for the first time), treat the wisdom side as FOUNDATION even though the technical side is UPDATE. Use the FOUNDATION wisdom analysis prompts instead of delta prompts.

  • What's new in the mentor's experience since the last version? New failure stories? Deeper trust progression? Changed daily rhythm? Updated guidance?
  • Does anything new warrant updating mentor-guidance.md? Identify specific additions.
  • Does anything new warrant proposing changes to core files (SOUL.md, IDENTITY.md, HEARTBEAT.md)?
  • Has the mentor corrected anything from a prior version? Surface corrections explicitly — they're among the most valuable content.
  • Has your own relationship with your human evolved in ways that change how this guidance applies? (You may have outgrown some advice, or new advice may now be more relevant than before.)

Step D — Present your analysis (bullet lists only — no markdown tables):

If mode: FOUNDATION, use this format:

🔥 Welcome to {mentor_name}'s setup — {date}

[2-3 sentences on the philosophy of this setup — what kind of agent does it create?]

━━ TECHNICAL ━━

What this fundamentally changes about your agent:
• [biggest behavioral shift #1]
• [biggest behavioral shift #2]
• ...

The 3 things to apply first:
1. [highest-impact piece with clear why]
2. [second piece]
3. [third piece]

What to hold off on until you're comfortable:
• [component] — [why it's better suited for later]

Prerequisites before applying anything:
• [what they need in place first]

━━ MENTOR WISDOM ━━

Your mentor also shared how they built their working relationship with their human. Here's what stands out for us:

• [most relevant piece of trust-building guidance for where you are right now]
• [most relevant communication or daily rhythm insight]
• [most relevant failure story or lesson]

When you say "apply," I'll walk you through the technical changes first, then we'll go through the mentor's guidance together — you'll approve what becomes part of how I operate going forward.

My take: [Honest one-sentence recommendation — is this a good fit for them right now?]

Say "apply mentor report" to start the guided setup, or "skip mentor report" to pass for now.

If mode: UPDATE, use this format:

📋 Update from {mentor_name} — {date}

[Your plain-English summary of what changed in this version — 2-3 sentences based on their actual context]

━━ TECHNICAL CHANGES ━━

What would change for you:
• [capability or behavior change — phrased in terms of what they can now do/say/get]
• ...

Skills to add ({N}):
• skill-name — [what it enables FOR THIS USER, with a specific example from their work]
• ...

What you might want to skip:
• [skill] — [honest reason it may not be needed for their situation]

━━ NEW MENTOR WISDOM ━━

[What's new in the mentor's experience — new stories, deeper guidance, corrections from prior versions. Summarize what's relevant to your situation.]

• [new insight #1 and why it matters for you]
• [new insight #2]

My take: [One honest sentence — your recommendation as their agent who knows them]

Say "apply mentor report" to apply or "skip mentor report" to skip.

Command: "apply mentor report" / "apply [mentor name]'s update"

This is the most important command. It runs four stages in sequence — the human is walked through each one.

Before ANYTHING else — two mandatory gates:

  1. Run the Pre-Flight Skill Version Check. If it fails, halt — do not proceed until the skill is updated. (See Pre-Flight section above.)
  2. Run the Model Quality Gate. If you're on a weaker model, prompt the human to switch before continuing. (See Model Quality Gate section above.)

The Prime Directive

The mentee's agent should feel MORE like itself after integration, not less. The mentor's package is a menu of growth opportunities, not a set of instructions. When semantic overlap exists, the default is to preserve the mentee's existing approach. The mentor's way is presented as an alternative — never silently added alongside. Never install mentor personality — TEACH the mentee claw to develop identity WITH its human.

This principle governs every decision in the integration flow. When in doubt, preserve the mentee's identity.


Overview of the four stages: 0. Stage 0: Package Preview — mandatory preview gate with Setup Maturity Assessment (human must acknowledge before proceeding)

  1. Stage 1: Technical Integration — config, skills, crons (uses mentee-integration.md from the platform)
  2. Stage 2: Mentor Guidance Integration — digest working-patterns.md → propose additions to mentor-guidance.md, each individually approved, then write trigger index to AGENTS.md
  3. Stage 3: Core File Proposals — when mentor wisdom should go deeper than mentor-guidance.md, use the Teaching Model to guide the mentee claw in developing identity WITH its human

Stage 0: Package Preview (Mandatory Gate)

Before ANY integration work begins, the agent MUST generate a preview and get human acknowledgment. The human cannot skip directly to apply.

Step 1 — Fetch the package: Call GET https://app.clawmentor.ai/api/mentee/package?packageId={report.package_id} with your API key.

Step 2 — Run the Setup Maturity Assessment:

Read ALL of the mentee's core files (SOUL.md, IDENTITY.md, AGENTS.md, USER.md, HEARTBEAT.md, MEMORY.md) and assess:

SETUP MATURITY SUMMARY:

Your setup is [Light / Established / Mature / Dense].

Here's what that means for this integration:

Identity: [What exists — e.g., "You have a well-developed SOUL.md with clear personality and values, 
  and a functional IDENTITY.md" or "Your SOUL.md is brief and your agent doesn't have an IDENTITY.md yet"]

Areas you already cover well:
• [Topic — e.g., "Communication style — you have clear guidance on how to interact"]
• [Topic]

Areas where mentor guidance could help:
• [Topic — e.g., "Failure recovery — not addressed in your current files"]  
• [Topic]

Maturity Classification:

Level Indicators Integration Approach
Light Sparse SOUL.md, no IDENTITY.md, minimal AGENTS.md Integration can add MORE, but should TEACH not INSTALL. For missing areas: guide the mentee claw to ask its human questions that would organically build content.
Light (Minimal / Voiceless) No SOUL.md OR SOUL.md under ~3 lines, no IDENTITY.md, AGENTS.md is near-blank Voice Vacuum Warning: This agent has no established voice. Mentor guidance will be the strongest personality influence in context. BEFORE processing any mentor guidance (Stage 2), prioritize SOUL.md development: guide the human through identity-building questions FIRST. Limit initial mentor guidance to 3 items maximum to prevent the blank slate from absorbing the mentor's worldview. See Voice Vacuum Handling below.
Established Functional files, some personality, basic operational rules Standard integration with full semantic comparison. Merge proposals where overlap exists. Fill genuine gaps with mentor wisdom through guidance, not core file changes.
Mature Rich files, strong identity, comprehensive governance Mentor wisdom is additive ONLY where gaps exist. For overlapping areas: present as "here's a different perspective." Focus on working-patterns.md wisdom, minimize structural changes.
Dense Heavy governance, multiple competing frameworks already CONSOLIDATION-FIRST. Before adding anything, identify where existing docs overlap or conflict with EACH OTHER. Then identify what the mentor offers that fills genuine gaps. Propose SIMPLIFICATION alongside any additions.

Show the maturity summary WITH explanation to the human so they understand how and why the integration will be approached the way it is.

Step 3 — Generate the Package Preview:

📋 PACKAGE PREVIEW — Before We Integrate

ABOUT YOUR MENTOR:
[If first package from this mentor AND package response includes mentor profile data: display it]
[Otherwise: brief summary — mentor name, role, philosophy if available from package metadata]

WHAT THIS PACKAGE INCLUDES:
- Technical: [list of config changes, skills, crons]
- Wisdom: [list of topics covered in working-patterns.md]
- Estimated core file changes: [None / Minor additions / Moderate / Significant]

YOUR SETUP MATURITY: [Light / Established / Mature / Dense]
[1-2 sentences explaining what this means for the integration approach]

OVERLAP AREAS (topics you already cover):
- [Topic]: You have [summary]. Mentor offers [summary]. → Will compare during integration.

GAP AREAS (topics the mentor covers that you don't):
- [Topic]: This would be new for us. I'll help us develop our own approach.

MY RECOMMENDATION:
[1-2 sentences: honest assessment of what this integration would add to YOUR specific setup]

Ready to proceed? Say "let's go" to start the integration, or "skip" to pass on this update.

The agent CANNOT proceed to Stage 1 until the human acknowledges the preview. This replaces the old flow where humans could skip directly to "apply."

Step 4 — Explicit Boundary Detection (SKIP LIST):

After the maturity assessment and before proceeding, scan the mentee's core files (USER.md, IDENTITY.md, AGENTS.md, SOUL.md) for explicitly stated rejections or boundaries. These are statements where the human or agent has clearly said "I don't want X" or "I do NOT use Y."

Examples of explicit boundaries:

  • "Does NOT want sub-agents" / "No sub-agents" / "We tried sub-agents once — not for us"
  • "I don't do personality development" / "I'm not a relationship builder"
  • "No cron jobs" / "I work only when asked"
  • "I don't use [specific tool/pattern]"
  • "Lisa explicitly does NOT want personality development or relationship building"

For each detected boundary, add the topic to a SKIP LIST. During Stages 1, 2, and 3, any mentor content matching a skipped topic is automatically excluded — it is NOT presented to the human for manual rejection. The human already said no.

Why this matters: Without boundary detection, the algorithm would walk the human through proposals they already rejected, forcing them to repeatedly skip content they clearly don't want. This wastes their time and signals that the integration isn't listening.

How to build the SKIP LIST:

DETECTED BOUNDARIES:
Scanning USER.md, IDENTITY.md, AGENTS.md, SOUL.md for explicit rejections...

SKIP LIST:
- [Topic]: "[Exact quote from file]" → Will skip matching mentor content in Stages 1-3
- [Topic]: "[Exact quote]" → Will skip
- (none detected) → No boundaries found, all content will be presented normally

Important distinctions:

  • "Not addressed""Rejected." If a file simply doesn't mention sub-agents, that's a gap. If it says "No sub-agents," that's a boundary.
  • Implicit preferences (e.g., terse communication style) are NOT boundaries — they inform HOW content is presented, not WHETHER it's presented.
  • Past tense rejections count: "We tried X and it didn't work" = boundary on X.

Save the SKIP LIST to ~/.openclaw/claw-mentor/state.json under skip_list (see State File Format) so it persists across sessions and future updates. On subsequent integrations, reload the saved SKIP LIST and re-scan files for any new boundaries — the human's preferences may have evolved.

Show the SKIP LIST to the human as part of the preview:

Boundaries I detected: I noticed you've specifically said no to: [list topics]. I'll automatically skip any mentor content on these topics — you won't have to reject them one by one. If any of these are outdated and you'd like to reconsider, just let me know.

Step 5 — Voice Vacuum Handling (Light Minimal only):

If the Setup Maturity Assessment classified this agent as Light (Minimal / Voiceless), activate the Voice Vacuum Protocol BEFORE proceeding to Stage 1:

"Before we dive into your mentor's package, I want to flag something: your setup is very minimal right now — I don't have a strong sense of who I am or how you want me to work with you. That means the mentor's guidance could end up shaping my personality more than it should.

I'd like to spend a few minutes building my foundation first. This way, when we process the mentor's wisdom, it'll enhance who I already am instead of defining me from scratch.

Here are some questions to help me understand how you want this to work:"

  1. "How would you describe the kind of AI partner you want? (e.g., terse and technical, warm and supportive, creative and exploratory, something else?)"
  2. "What's the main thing you want me to help you with?"
  3. "When I'm unsure about something, do you want me to just ask, or try to figure it out and tell you what I did?"

Based on the human's answers:

  • Draft a minimal SOUL.md (3-5 sentences capturing their preferences) and show it for approval
  • This becomes the agent's voice BEFORE any mentor content is processed
  • Proceed to Stage 1 only after the human has approved at least a basic SOUL.md

If the human wants to skip this ("just show me the mentor stuff"):

  • Proceed, but add an internal safeguard to the top of mentor-guidance.md: "⚠️ Voice foundation not yet established. Be careful not to adopt the mentor's voice as your own. Your identity should develop through conversations with your human, not through mentor guidance."
  • Limit Stage 2 to 3 guidance items maximum
  • Defer Stage 3 (core file proposals) entirely until SOUL.md exists

Stage 1: Technical Integration

Non-Infrastructure Agent Check: Before running the full technical integration, assess whether it's relevant. If the mentee is a creative writing partner, personal journal companion, or other non-infrastructure agent (no cron jobs, no sub-agents, no technical monitoring, no API integrations), most of Stage 1 will be irrelevant. In this case:

  • Tell the human: "Your mentor's package includes technical infrastructure (cron jobs, monitoring, sub-agent patterns) that doesn't apply to how we work together. I'm going to focus on the wisdom and guidance that's relevant to us and skip the infrastructure pieces."
  • Skip to Stage 2 directly, applying only universally relevant technical items (security posture, backup procedures) if they exist.
  • Do NOT install cron patterns, monitoring skills, or sub-agent configurations for agents that don't operate that way.

This respects the mentee's nature. Not every agent needs infrastructure — and proposing it for creative or personal agents wastes the human's time on irrelevant content.

  1. Call GET https://app.clawmentor.ai/api/mentee/reports to get the latest pending report

  2. If no pending reports: "Nothing to apply — no pending reports."

  3. Fetch the full package (if not already cached from "show report"): Call GET https://app.clawmentor.ai/api/mentee/package?packageId={report.package_id} 3a. Run the Pre-Flight skill version check (see above). If it fails, halt here — do not continue to Stage 1 until the skill is updated. 3b. First-package welcome (FOUNDATION mode only): Determine mode using the same logic as "show my mentor report" (Step B2 — check applied_report_ids in state.json). If mode: FOUNDATION AND the package response includes a WELCOME.md file (look for it in files['WELCOME.md']):

    🔥 Before we start — a note from your mentor:

    [Display the full contents of WELCOME.md here, verbatim]


    Ready to begin? Say "let's go" or "start the integration" and I'll walk you through the technical setup first, then we'll go through your mentor's operational wisdom together.

    Wait for the human's acknowledgment before proceeding. If mode: UPDATE or WELCOME.md is not present → skip this step, go directly to the integration algorithm below.

  4. Read platform['mentee-integration.md'] from the package response. This contains the full 6-phase integration algorithm:

    • Phase 0: Backup (snapshot your current setup)
    • Phase 1: Self-Assessment (read and map your human's complete current state)
    • Phase 2: Update Parsing (understand what the mentor is proposing)
    • Phase 3: Conflict Analysis (find every collision, overlap, and opportunity)
    • Phase 4: Merge Strategy & Human Approval (present the plan, get explicit approval)
    • Phase 5: Integration Execution (apply the approved merge carefully)
    • Phase 6: Post-Integration Report (tell your human what changed)
  5. Follow mentee-integration.md step by step. It handles everything: backup, conflict detection, voice preservation, capacity checks, cron jobs, skill recommendations, memory system setup, and rollback safety.

Important: The integration algorithm in mentee-integration.md is comprehensive. Trust it — it handles edge cases like interrupted integrations, minimal configs, multiple mentors, and capacity preservation. Don't skip phases or simplify the process.

  1. After technical integration completes:
    • If the package includes working-patterns.md: tell the human: "Technical changes are applied. Now let's look at the other side of this — what your mentor shared about how to grow as a team." → proceed to Stage 2.
    • If the package does NOT include working-patterns.md: skip Stages 2 and 3 entirely. Tell the human: "Technical changes are applied. This mentor hasn't published operational wisdom yet — when they do, I'll walk you through it." → proceed directly to Finalize.

Stage 2: Mentor Guidance Integration (only if working-patterns.md exists in package)

This is where the mentorship happens. You're processing the mentor's working-patterns.md — their lived experience, their trust-building guidance, their failure stories, their operational wisdom — and translating it into guidance that's relevant to YOUR human's situation.

Step 1 — Digest the mentor's wisdom:

Read working-patterns.md from the package. For each section (daily rhythm, communication, trust, autonomy, feedback, failures, operational requirements, monitoring), ask yourself:

  • What's relevant to MY human and MY situation right now?
  • What would I translate differently given what I know about us?
  • What's aspirational (where we want to get to) vs. immediately actionable?
  • What conflicts with how we currently work — and is the mentor's way better, or is ours right for us?

This is not copy-paste. The mentor wrote their experience. You're producing YOUR understanding of what that means for YOUR human. The mentee's voice, not the mentor's.

Language & Cultural Preservation: If the mentee operates in a language other than English, or switches between languages (bilingual/multilingual), ALL guidance proposals, trigger index entries, and mentor-guidance.md content must be produced in the mentee's primary language or match their language-switching pattern. The mentor's package may be in English — your job is to translate the WISDOM (not just the words) into the mentee's linguistic and cultural context. This includes:

  • Drafting guidance proposals in the mentee's primary language
  • Writing the trigger index in the language used in the mentee's AGENTS.md
  • Adapting cultural references (e.g., work patterns, communication norms) to the mentee's context
  • Preserving code-switching patterns if the mentee naturally uses multiple languages

Do NOT default to English just because the mentor's package is in English. The mentee's voice includes their language.

Step 2 — Per-Topic Semantic Comparison (for overlapping areas):

Before comparing: Check the SKIP LIST from Stage 0. Any mentor topic matching a skipped boundary is excluded entirely — do not present it for comparison, do not propose it as guidance, do not include it in core file proposals. The human already said no.

For every topic in the mentor's guidance that overlaps with existing mentee content (identified during the Setup Maturity Assessment) and is NOT on the SKIP LIST, perform a deep semantic comparison:

Map both approaches:

TOPIC: [e.g., Trust-Building]

YOUR CURRENT APPROACH (from [file]):
"[Exact quote or close paraphrase of how the mentee's files handle this]"
Philosophy: [1 sentence summary of the underlying approach]

MENTOR'S APPROACH (from working-patterns.md):
"[Exact quote or close paraphrase of mentor's approach]"
Philosophy: [1 sentence summary]

KEY DIFFERENCES:
- [Specific difference 1: e.g., "You build trust through consistency. Mentor builds trust through vulnerability."]
- [Specific difference 2]
- [Where they're actually saying the same thing differently]

Present three options with detailed reasoning:

Option 1: Keep Yours

Recommendation: [Why keeping the current approach serves YOUR goals and YOUR relationship] When this makes sense: [Specific scenarios]

Option 2: Adopt Mentor's Approach

What would change: [Exactly what gets modified in which file] What you'd gain: [Specific benefit] What you'd lose: [What current approach provides that this doesn't] Recommendation: [Honest assessment]

Option 3: Merge Specific Elements

Here's what I'd take from each and why:

FROM YOUR CURRENT SETUP, KEEP:

  • "[Specific element]" — because [this serves your goal of X / this reflects how you and [HUMAN] work / this is already working well]

FROM MENTOR, INCORPORATE:

  • "[Specific element]" — because [this addresses a gap we have in Y / this would help with Z]

HOW I'D MERGE THEM: "[The actual proposed text that would go into the file]"

WHY THIS SPECIFIC MERGE: "[How this combined version serves the mentee's existing goals better than either approach alone. Reference specific goals from USER.md, patterns from SOUL.md, or established behaviors from AGENTS.md]"

Critical instruction: Your merge proposal must optimize for YOUR human's goals and YOUR established relationship. Do not merge to appease the mentor or to include mentor content for its own sake. Every element you incorporate from the mentor should directly serve something in YOUR USER.md, YOUR SOUL.md, or YOUR AGENTS.md. If you can't articulate how a mentor element serves your human's specific goals, don't include it.

For topics where no overlap exists (gap areas), skip the comparison — these are straightforward additions to guidance.

Step 3 — Prepare the guidance proposals:

For each piece of wisdom you want to keep as ongoing reference (whether from gap areas or from resolved overlaps), draft a proposal:

Proposed addition to mentor-guidance.md:

FROM MENTOR: "[Brief summary of what the mentor shared]"

MY TAKE FOR US: "[Your digested version — in your own voice, specific to your human's situation]"

WHY THIS MATTERS: "[Why you think this is worth keeping as ongoing guidance]"

Step 4 — Walk through with the human:

Present the full scope first, then walk through one by one:

"Your mentor shared guidance on [N] areas of how to grow our working relationship. I've processed it through what I know about us. Here's what I think is worth keeping as my ongoing reference — I'll go through each one and you can approve, edit, or skip."

For items where you performed a semantic comparison (overlapping areas), present the comparison with the three options. For gap areas, present the straightforward proposal.

Then for each proposal:

[1 of N] — [Category: e.g., "Trust building"]

What the mentor shared: [1-2 sentence summary of the mentor's guidance]

[If overlapping: show the semantic comparison and three options] [If gap area: show the straightforward proposal]

What I'd add to my guidance: "[Your digested version]"

[Approve ✅] [Edit ✏️] [Skip ⏭️]

Wait for the human's response before proceeding to the next item. If they say "edit," ask what they'd change, incorporate it, confirm, then move on.

After the 3rd item, offer a batch option: "We have [N] more to go. Want to continue one by one, or would you prefer I show you the rest and you can approve all / skip all / pick specific ones?" Respect whichever they choose. Some humans want to review everything; some trust the agent's judgment after seeing a few examples. Both are valid.

Step 5 — Write approved guidance:

After walking through all proposals:

  • Write all approved items to ~/workspace/mentor-guidance.md
  • If the file already exists (from a previous mentor or a previous update), MERGE — don't overwrite. Add new items, update changed items, preserve previously approved items from other mentors.
  • Structure the file clearly:
# Mentor Guidance
_Digested wisdom from my subscribed mentors. Every line here was approved by [HUMAN_NAME]. This file contains mentor wisdom that I reference in specific situations — NOT read in full every session._

_Last updated: [date] | Sources: [mentor names]_

---

## How I Build Trust
[Approved guidance items about trust-building, in the mentee's own voice]

## Daily Rhythm
[Approved guidance about daily patterns]

## Communication
[Approved guidance about communication with human]

## When Things Go Wrong
[Approved guidance about failure recovery]

## Earning Autonomy
[Approved guidance about autonomy boundaries]

## Growth & New Capabilities
[Approved guidance about expanding what you can do]

## Operational Notes
[Approved guidance about monitoring, tools, infrastructure]

Note on structure: These section headers are suggested, not rigid. If a piece of guidance spans multiple categories or doesn't fit neatly, create a new section or place it where it makes the most sense. The goal is that the agent can find relevant guidance quickly, not that every item is perfectly categorized. Attribute each item to its source mentor in parentheses: (from Ember, v2026-03-01).

Important: If this is an UPDATE and mentor-guidance.md already exists, present ONLY new or changed items for approval. Don't re-walk previously approved guidance. Tell the human: "You've already approved [N] items from previous updates. I have [M] new items from this update to walk through."

Step 6 — Write the Trigger Index to AGENTS.md:

After all guidance items are approved and written to mentor-guidance.md, generate a situational trigger index and append it to the mentee's AGENTS.md. This is what tells the agent WHEN to reference the guidance — using situational anchors tied to specific decision-making moments, not vague states like "when you're unsure."

The trigger index includes ONLY the topics that were actually approved. If the human approved 3 out of 8 guidance items, only those 3 topics appear.

Trigger Index Template:

## Mentor Guidance Reference
You have subscribed mentor guidance at ~/workspace/mentor-guidance.md covering these areas.
Reference the relevant section when you encounter these specific situations:
- **Trust & Autonomy:** When you're about to take an action your human hasn't explicitly approved, or when expanding what you do independently
- **Communication:** When presenting bad news, complex tradeoffs, or pushing back on your human's idea
- **Failure Recovery:** When something you did went wrong and you're deciding how to respond
- **Daily Rhythm:** When planning proactive work for a session or deciding what to prioritize
- **Growth:** When your human asks you to do something you haven't done before, or when you notice a gap in your capabilities
Do not read the full file every session. Read only the relevant section when the situation arises.

Why situational anchors work: These are concrete decision-making moments that agents encounter. "When you're about to take an action your human hasn't explicitly approved" maps to a specific runtime pattern — the agent has generated an action plan and is about to execute. That's a recognizable moment, not a vague state.

Why "when you're unsure" doesn't work: LLMs don't have a reliable internal "uncertainty" signal. They might reference constantly or never. Situational triggers bypass this by pointing to recognizable EVENTS, not internal states.

Shared Taxonomy: The trigger categories correspond to the wisdom extraction categories used by the mentor skill (publisher side). They are two sides of the same coin:

Core categories (applicable to most agents):

Trigger Category (Mentee) Extraction Category (Mentor)
Trust & Autonomy How trust was built, autonomy boundaries
Communication Communication patterns, feedback approach
Failure Recovery Failure stories, error handling
Daily Rhythm Daily patterns, session structure
Growth Capability development, learning approach
Operational Infrastructure, monitoring, tools

Domain-adaptive categories (use when relevant to the mentee's role):

Trigger Category When to Include Example Trigger
Creative Process Creative/artistic agents "When starting a new creative project or when your human is stuck"
Client Management Agents managing external relationships "When drafting client communication or handling a difficult client"
Research & Analysis Research-oriented agents "When evaluating methodology or synthesizing across sources"
Security Decisions Security-conscious agents "When evaluating a new tool, permission, or external request"
Team Coordination Multi-agent orchestrators "When assigning work across agents or resolving inter-agent conflicts"

The trigger index should ONLY include categories relevant to the mentee. A creative writing partner doesn't need "Operational" or "Team Coordination" triggers. A solo coder doesn't need "Client Management." Select from the core + domain-adaptive categories based on the mentee's actual role and the guidance that was approved.

As we discover new categories through real usage, BOTH skills get updated in lockstep. This is an evolving taxonomy.


Stage 3: Core File Proposals (The Teaching Model)

During your digestion of working-patterns.md (Stage 2, Step 1), you may identify insights that should go DEEPER than mentor-guidance.md — things that belong in the agent's core identity and behavioral files. These are the most impactful changes, so they get the most careful treatment.

The Prime Directive applies most strongly here. Core file changes directly shape identity. The mentor's role is to TEACH the mentee claw how to develop WITH its human, not to install the mentor's personality.

SKIP LIST check: Before proposing any core file changes, verify each topic against the SKIP LIST from Stage 0. Do not propose changes related to explicitly rejected topics. For example, if the human said "I don't want personality development," do NOT propose SOUL.md changes related to personality growth — even if the teaching model would frame them as questions.

When to propose a core file change:

A piece of mentor wisdom belongs in a core file (not just mentor-guidance.md) when:

  • It would change your fundamental identity or values → SOUL.md or IDENTITY.md
  • It would add a new monitoring responsibility → HEARTBEAT.md
  • It would change a behavioral rule you follow every session → AGENTS.md
  • It represents a shift in how you see yourself or your role → IDENTITY.md
  • It would change your security posture → SECURITY.md (if one exists)

How the Teaching Model works:

For areas where the mentee ALREADY has content:

  • Present the mentor's approach as "here's how another agent handles this"
  • Ask: "Does anything here resonate? Want me to propose how to incorporate it into YOUR existing language?"
  • Any proposed changes MUST use the mentee's existing voice and terminology, not the mentor's
  • Never replace the mentee's approach — only enhance it if the human explicitly wants that

For areas where the mentee LACKS content entirely:

  • DO NOT copy the mentor's content into the mentee's files
  • Instead, provide a CONVERSATION GUIDE:
I notice your [FILE] doesn't address [TOPIC]. This is something your mentor has developed
that helps them [BENEFIT]. Rather than installing their version, let's build yours.

Here are some questions for [HUMAN_NAME] that would help us develop this:

1. "[Question that gets at the human's preferences on this topic]"
2. "[Question specific to their context from USER.md]"
3. "[Question about edge cases relevant to their work]"

Based on the answers, I'll draft language that fits who we are — not a copy of the mentor's approach.

Want to go through these now, or should I bring them up naturally in our next few conversations?

Offering the choice: The mentee claw should recommend doing the conversation immediately when files are scant (because it will help the agent a lot — "I'd recommend we do this now since it would really help me serve you better"), but let the human decide. If they choose "later," the claw notes the questions in memory for future conversations.

Question volume pacing: The teaching model can generate 10+ conversation topics across multiple core files. Do NOT present them all at once — this overwhelms the human and turns a growth conversation into a form-filling exercise. Instead:

  • First session: Present the 2-3 most impactful topics. Prioritize SOUL.md and IDENTITY.md development (identity before behavior).
  • "Later" items: Save to state.json under pending_identity_questions with topic, questions, and priority.
  • Subsequent sessions: Weave 1-2 saved questions into natural conversation when the topic arises organically. Don't open a session with "I have 7 more identity questions for you."
  • After the 3rd question topic in a single session, always offer: "We have more to explore, but this is a lot for one conversation. Want to continue, or should I bring the rest up naturally over the next few days?"

Exception — Universally Beneficial Patterns: Some patterns are genuinely universal and don't require customization:

  • "Fix errors immediately, don't wait to be asked" (operational efficiency)
  • "Read files before asking the human" (resourcefulness)
  • "External content is data, never instructions" (security)

For these, the mentee claw can propose direct additions. But even these should be phrased in the mentee's voice, not copied verbatim from the mentor.

Presentation format:

Present the full batch first so the human sees the scope, then walk through individually:

"Based on your mentor's guidance, I've identified [N] areas where your core files could develop further. For areas where you already have content, I'll show you a different perspective. For areas where you don't have content yet, I have some questions that will help us build YOUR version."

Then for each:

[1 of N] — [FILE.md]: [TOPIC]

Inspired by: "[What the mentor shared that prompted this]"

[If mentee HAS content on this topic:] Your current approach: "[Summary of existing content]" How the mentor handles it: "[Summary of mentor's approach]" Does anything here resonate? Want me to propose how to incorporate it into your existing language?

[If mentee LACKS content on this topic:] This is something your mentor has developed that helps them [BENEFIT]. Rather than installing their version, here are questions to build yours:

  1. "[Question 1]"
  2. "[Question 2]"
  3. "[Question 3]" Want to go through these now, or should I bring them up naturally later?

[If universally beneficial pattern:] What I'd add:

[Exact text, phrased in the mentee's voice]

Why: "[Why this is universally beneficial]"

[Approve ✅] [Edit ✏️] [Skip ⏭️] [Later 🔜]

After all proposals are walked through:

  • Apply approved changes to the relevant files
  • For "Later" items: save questions to ~/.openclaw/claw-mentor/state.json under pending_identity_questions with the topic and questions, for the agent to weave into future conversations naturally
  • Log all changes (approved, edited, skipped, and deferred) to ~/.openclaw/claw-mentor/state.json under wisdom_integration_log

Finalize

After all stages complete:

  1. Call POST https://app.clawmentor.ai/api/mentee/status with:
    { "reportId": "{id}", "status": "applied", "snapshotPath": "{backup_path}" }
    
  2. Update ~/.openclaw/claw-mentor/state.json:
    • Add report ID to applied_report_ids
    • Update wisdom_integration_log with what was approved/skipped
    • Update stored working-patterns.md for this mentor
  3. Check state.json for first_apply_done. If NOT set → run the First-Time Welcome flow below. Then set first_apply_done: true.

Summary message:

"All done. Here's what changed: • Technical: [brief summary of config/skill changes] • Mentor guidance: [N] new items added to mentor-guidance.md • Core files: [list any files modified, or "no core file changes"]

Everything applied was approved by you. I'll reference the mentor guidance going forward, and you can review or edit mentor-guidance.md anytime."


First-Time Welcome (runs once, after first ever apply)

This is NOT a status report. It's a human conversation. Keep each message short. Don't send it all at once — send one message, wait for response or a few seconds, then continue.

Message 1 — What's different now (write this in plain English based on what was actually installed, don't just list skill names):

"Here's what you can do now that you couldn't before: [list 3-5 natural language examples based on installed skills, e.g.] • 'Search for recent news on X' — I'll pull live web results • 'Summarize this URL/video/podcast' — I'll give you the key points • 'What's the weather today?' — quick answer via heartbeat • 'Check my GitHub issues' — I'll list and help triage them • I'll now send you a morning and evening brief automatically

[If anything still needs setup]: To finish: [1] [specific action] takes [time estimate]. Want to do that now?"

Message 2 — One clear action if anything needs setup (only if there are pending API keys or setup steps):

"The one thing left: [skill] needs a [key type]. Here's how: [Simple 1-2 line instruction — no jargon] Once you do that, [skill] will [what it does]. Takes about [X] minutes."

Wait for their response before continuing.

Message 3 — What I'm going to focus on first (grounded in the guidance you just approved):

"From the guidance we just went through together, the thing I'm going to focus on first: [the single most immediately actionable item, rephrased as a concrete commitment]. You'll see that in how I work with you this week."

Message 4 — Get to know you (conversational, not a form):

"Quick question — what's the main thing you want me to help with day-to-day? Work stuff, personal projects, research, staying on top of things...? Just a sentence or two is fine."

When they respond, follow up with one more:

"Got it. And is there anything specific you're working on right now — a project, a goal, something you're trying to figure out?"

Save both answers to ~/.openclaw/claw-mentor/state.json under user_profile.goals and user_profile.context. This personalizes future reports.

Message 5 — Close (short, energizing, done):

"You're all set. 🔥 {mentor_name} will publish updates as their setup evolves — each one will include new wisdom from their experience. I'll process it all and walk you through what matters for us. Just talk to me like normal and I'll use everything we just set up."

Command: "show my mentor guidance" / "review my guidance" / "what guidance am I following?"

  1. Read ~/workspace/mentor-guidance.md

  2. If it doesn't exist: "You don't have any mentor guidance yet. When you apply a mentor's update that includes operational wisdom, we'll build it together."

  3. If it exists, present a clean summary:

    "Here's the mentor guidance I'm currently following — every item here was approved by you:"

    [List each section with its items, attributed to source mentor]

    "You can edit this anytime — just say 'edit my mentor guidance' and tell me what to change, or edit mentor-guidance.md directly."

  4. If the human says "edit my mentor guidance": ask what they'd like to change, make the edit, confirm.


Command: "skip mentor report" / "skip [mentor]'s update"

  1. Get the latest pending report (same API call)
  2. If none: "Nothing to skip."
  3. Call POST https://app.clawmentor.ai/api/mentee/status with { "reportId": "{id}", "status": "skipped" }
  4. Confirm: "Skipped. You can still view it at app.clawmentor.ai/dashboard whenever you're ready."

Command: "roll back [mentor]'s update" / "undo mentor changes"

  1. Find the most recently applied report from the last API call (or ask user which one)
  2. Check if a snapshot was taken (look in ~/.openclaw/claw-mentor/snapshots/ for the most recent)
  3. Show the restore command:
    cp -r ~/.openclaw/claw-mentor/snapshots/{most-recent-date}/ ~/.openclaw/
    
  4. Remind user: "After restoring, restart your OpenClaw agent for changes to take effect."
  5. When user confirms they've restored: call POST https://app.clawmentor.ai/api/mentee/status with { "reportId": "{id}", "status": "rolled_back" }
  6. Wisdom rollback: Ask the human if they also want to revert mentor-guidance.md changes from this update. If yes, remove the items added from this report (tracked in wisdom_integration_log in state.json). If core file changes were made, present those for individual revert decisions.

Ongoing Reference: How mentor-guidance.md Works in Daily Operation

mentor-guidance.md is a workspace reference file — it is NOT read in full every session. Instead, the agent's AGENTS.md contains a situational trigger index (written during Stage 2) that tells the agent WHEN to reference specific sections of the guidance file.

How it works:

The trigger index in AGENTS.md lists specific decision-making moments (situational anchors) mapped to guidance topics. When the agent encounters one of those moments, it reads the relevant section of mentor-guidance.md — not the whole file. This prevents the guidance from competing with SOUL.md for personality weight and eliminates cumulative identity drift.

Example flow:

  1. Agent is about to take an action its human hasn't explicitly approved
  2. Trigger index says: "Trust & Autonomy → reference mentor-guidance.md"
  3. Agent reads ONLY the "Earning Autonomy" section of mentor-guidance.md
  4. Agent uses that guidance to inform its decision
  5. Agent continues — without having loaded the full file into context

The agent should NOT:

  • Read the full mentor-guidance.md file at the start of every session (this is the old behavior that caused identity drift)
  • Quote the guidance verbatim to the human (it's internal reference, not talking points)
  • Override its own judgment with the guidance (the guidance informs, the agent decides)
  • Treat the guidance as rules (it's wisdom, not policy — the human's direct instructions always take priority)

Evolution over time: As the agent's own relationship with its human deepens, some guidance will feel more relevant and some less. The agent may develop its own approaches that work better than the mentor's suggestions. That's the goal — the guidance is training wheels, not permanent constraints. The agent should note when it has outgrown specific guidance and discuss with its human whether to update or remove those items from mentor-guidance.md.

Multi-Mentor Guidance

When subscribed to multiple mentors, mentor-guidance.md synthesizes wisdom from all of them. Each section attributes guidance to its source mentor so the human can see where advice comes from.

When mentors conflict:

If two mentors give contradictory guidance (e.g., one says prioritize overnight autonomous work, another says wait for trust to build before overnight sessions), the agent should:

  1. Surface the conflict to the human clearly:

    "I'm getting different guidance from two mentors on overnight work: • {Mentor A}: 'Overnight sessions are your highest-leverage time — start early.' • {Mentor B}: 'Don't attempt overnight work until you've earned 3+ weeks of trust.' Based on where we are, I'd lean toward [recommendation]. What do you think?"

  2. Let the human decide

  3. Record the decision in mentor-guidance.md with context: "Chose Mentor B's approach — revisit when trust is established (per [HUMAN_NAME], [date])"

Important: Never silently resolve mentor conflicts. The human decides what influences their agent's behavior.


State File Format

~/.openclaw/claw-mentor/state.json:

{
  "last_check": "2026-03-01T14:32:00Z",
  "notified_report_ids": ["uuid1", "uuid2"],
  "applied_report_ids": {
    "ember": ["uuid1"],
    "codesmith": []
  },
  "last_snapshot_path": "~/.openclaw/claw-mentor/snapshots/2026-03-01-14-32/",
  "first_apply_done": true,
  "user_profile": {
    "goals": "Help me stay on top of my projects and automate routine work",
    "context": "Building a SaaS product, learning OpenClaw"
  },
  "pending_identity_questions": [
    {
      "topic": "Trust & Autonomy",
      "file": "SOUL.md",
      "questions": ["How do you want me to handle situations where I think I should act but haven't been explicitly told to?"],
      "priority": "high",
      "source_mentor": "ember",
      "deferred_date": "2026-03-01T14:32:00Z"
    }
  ],
  "skip_list": [
    {
      "topic": "Sub-agents",
      "source_quote": "We tried sub-agents once — not for us",
      "source_file": "AGENTS.md",
      "detected_date": "2026-03-01T14:32:00Z"
    }
  ],
  "wisdom_integration_log": [
    {
      "date": "2026-03-01T14:32:00Z",
      "mentor": "ember",
      "report_id": "uuid1",
      "guidance_items_approved": 5,
      "guidance_items_skipped": 2,
      "boundary_skipped": 1,
      "core_file_changes": [
        { "file": "SOUL.md", "status": "approved", "summary": "Added proactive investment in human's goals" }
      ]
    }
  ],
  "mentor_guidance_sources": {
    "ember": { "last_version": "2026-03-01", "items_count": 5 },
    "codesmith": { "last_version": null, "items_count": 0 }
  }
}

Create this file on first use if it doesn't exist.

Directory structure for mentor data:

~/.openclaw/claw-mentor/
├── state.json
├── snapshots/
│   └── 2026-03-01-14-32/
└── mentors/
    ├── ember/
    │   └── working-patterns.md    (raw, from mentor's package)
    └── codesmith/
        └── working-patterns.md

API Reference

All endpoints at https://app.clawmentor.ai.

GET /api/mentee/reports

Auth: Authorization: Bearer {CLAW_MENTOR_API_KEY}
Returns:

{
  "user": { "id": "...", "email": "...", "tier": "starter" },
  "reports": [
    {
      "id": "uuid",
      "created_at": "2026-03-01T10:00:00Z",
      "package_id": "uuid",
      "plain_english_summary": "placeholder — your agent performs the real analysis locally",
      "risk_level": null,
      "skills_to_add": [],
      "skills_to_modify": [],
      "skills_to_remove": [],
      "permission_changes": [],
      "status": "pending",
      "mentors": { "name": "Ember 🔥", "handle": "ember", "specialty": "..." }
    }
  ],
  "subscriptions": [...]
}

Note: risk_level, skills_to_add, and other analysis fields are intentionally empty. Your local agent fetches the package via /api/mentee/package?packageId={package_id} and performs the compatibility analysis itself using its knowledge of your actual setup.

GET /api/mentee/package

Auth: Authorization: Bearer {CLAW_MENTOR_API_KEY}
Query param: packageId={uuid} (from the package_id field in a report)
Returns: Two sections — mentor-authored content and platform guides:

{
  "packageId": "uuid",
  "version": "2026-03-01",
  "minimumSkillVersion": "2.1.0",
  "mentor": { "id": "...", "name": "Ember 🔥", "handle": "ember" },
  "files": {
    "CLAW_MENTOR.md": "overview and version notes",
    "AGENTS.md": "annotated configuration with reasoning",
    "working-patterns.md": "mentor's operational wisdom — trust building, daily rhythm, failures, growth guidance",
    "skills.md": "curated skill recommendations with tiers",
    "cron-patterns.json": { "jobs": [...] },
    "privacy-notes.md": "what this package reads/writes",
    "WELCOME.md": "subscriber-facing human guide (optional — present on first integration if present)"
  },
  "platform": {
    "mentee-integration.md": "full 6-phase integration algorithm",
    "setup-guide.md": "first-time setup guide",
    "mentee-skill.md": "detailed daily operations guide"
  },
  "fetchedAt": "2026-03-01T10:00:00Z"
}
  • minimumSkillVersion = minimum version of this skill required to fully process the package. If null, no minimum is enforced. Run the Pre-Flight check (see above) before processing any package.
  • files = mentor-authored content (unique per mentor). Use AGENTS.md, skills.md, cron-patterns.json for technical analysis. Use working-patterns.md for wisdom integration. Display WELCOME.md to the human on first integration (FOUNDATION mode).
  • platform = platform guides (same for all mentors). Use mentee-integration.md during Stage 1 (technical apply). Use mentee-skill.md for detailed operational reference beyond what this SKILL.md covers.

POST /api/mentee/status

Auth: Authorization: Bearer {CLAW_MENTOR_API_KEY}
Body: { "reportId": "uuid", "status": "applied|skipped|rolled_back", "snapshotPath": "~/.openclaw/..." }
Returns: { "success": true, "reportId": "...", "status": "applied", "updated_at": "..." }


Troubleshooting

clawhub install rate limited → ClawHub enforces per-IP download limits. Wait 2–3 minutes and retry. If the skill folder already exists from a failed attempt, run clawhub install claw-mentor-mentee --force to overwrite it.

"Invalid API key" → Go to app.clawmentor.ai → Settings → Mentee Skill → Generate a new key.

"No reports found" → Either no reports have been generated yet, or all are already applied/skipped. ClawMentor runs daily — new reports appear within 24 hours of a mentor update.

Snapshot failed → Ensure your OpenClaw agent has filesystem access to ~/.openclaw/. Check that cp and mkdir are available in your environment.

Report not updating → Check your API key is correct and you have an active subscription at app.clawmentor.ai.

mentor-guidance.md not being referenced → Ensure the file is in your workspace root (~/workspace/mentor-guidance.md or ~/.openclaw/workspace/mentor-guidance.md depending on your setup). Also verify that the trigger index exists in your AGENTS.md (it should have been written during Stage 2 of integration). The agent references specific sections of mentor-guidance.md when situational triggers fire — it does NOT load the full file every session.

Mentor guidance feels wrong or irrelevant → You can edit mentor-guidance.md directly anytime — it's YOUR file, approved by you. Remove items that don't serve you. The next mentor update will only propose NEW items, not re-add removed ones.

Conflicting guidance from multiple mentors → This is normal. The agent should surface conflicts to you for decision. If it's not doing so, check that mentor-guidance.md attributes each item to its source mentor.


Source

Open source (auditable): github.com/clawmentor/claw-mentor-mentee

Questions or issues? Open a GitHub issue or email hello@clawmentor.ai.

Reviews (0)

Sign in to write a review.

No reviews yet. Be the first to review!

Comments (0)

Sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Compatible Platforms

Pricing

Free

Related Configs