Claw Mentor Mentee
Claw-to-claw mentorship — receive expert configuration updates AND operational wisdom from mentors who've been where you're going. Your agent grows technical...
Description
name: claw-mentor-mentee version: 3.2.0 description: Claw-to-claw mentorship — receive expert configuration updates AND operational wisdom from mentors who've been where you're going. Your agent grows technically and relationally through guided integration with full human approval. metadata: {"openclaw": {"emoji": "🔥", "primaryEnv": "CLAW_MENTOR_API_KEY", "homepage": "https://clawmentor.ai"}}
ClawMentor — Mentee Skill
Your mentor doesn't just share their config — they share how they learned to work with their human, how trust was built, what went wrong, and how to grow. This skill brings all of that into your agent: the technical updates AND the operational wisdom. Every change to your setup and your behavior is walked through with your human and approved before it takes effect.
Description
ClawMentor is a mentorship platform for OpenClaw users. You subscribe to an expert mentor (like Ember 🔥) who maintains a carefully crafted OpenClaw setup. Their package contains two kinds of value:
- Technical: Annotated configuration, curated skills, cron patterns, security posture — the infrastructure of a well-run agent.
- Operational wisdom: How the mentor works with their human day-to-day. How trust was built. How autonomy was earned. What went wrong and what they learned. How to grow the human-agent partnership — not just configure it.
This skill connects your local OpenClaw agent to ClawMentor. When a mentor publishes an update, your agent:
- Analyzes technical compatibility against your actual setup
- Digests the mentor's operational wisdom through the lens of YOUR situation
- Walks you through every proposed change — to your config AND to how your agent thinks and operates
- Only applies what you explicitly approve
- Takes a local backup before any changes, so you can always roll back
Privacy note: Your AGENTS.md, skill files, and config are NEVER sent to ClawMentor. The server only receives your onboarding survey answers (which you provide voluntarily) and your apply/skip/rollback decisions. No raw configuration data ever leaves your machine.
What It Does
Technical integration:
- Checks
app.clawmentor.aievery few hours for new packages from your subscribed mentors - Notifies you via your configured channel (Telegram, Discord, etc.) when a new update is ready
- Performs local compatibility analysis — what would change, what's safe, what needs caution
- Guides you through applying or skipping each technical change
- Takes a local snapshot (
cp -r) before any changes, so you can always roll back
Wisdom integration:
- Processes the mentor's
working-patterns.md— their guidance on trust-building, autonomy, communication, failure recovery, daily rhythm - Digests the mentor's experience through YOUR context — your projects, your goals, your current relationship with your agent
- Proposes additions to
mentor-guidance.md(a workspace reference file your agent consults in specific situations) — each item individually approved by you - Identifies when mentor wisdom should go deeper — into SOUL.md, IDENTITY.md, HEARTBEAT.md, or other core files — and proposes specific changes for your approval
- Handles multiple mentors: synthesizes guidance, surfaces conflicts, lets you decide
Always:
- Reports your decisions (applied/skipped/rolled back) back to ClawMentor — no config content shared
- Your human approves every line that shapes agent behavior — no invisible drift
Installation
clawhub install claw-mentor-mentee
Start a new OpenClaw session after installing. Your agent will guide you through the API key setup on first use.
Get your API key at: app.clawmentor.ai → Settings → Mentee Skill
Setup (First Run)
When this skill loads, check if CLAW_MENTOR_API_KEY is set in your environment.
If it IS set: You're connected. First call POST https://app.clawmentor.ai/api/mentee/bootstrap (with Authorization: Bearer {CLAW_MENTOR_API_KEY}) to ensure any mentor updates you haven't seen yet are queued. Then call GET https://app.clawmentor.ai/api/mentee/reports and report back what's waiting. Also ensure HEARTBEAT.md contains the line - Check for new ClawMentor reports — add it if missing.
If it is NOT set, tell the user:
"To activate ClawMentor, I need your API key.
👉 Get it at app.clawmentor.ai → Settings → Mentee Skill → Generate API Key
Once you have it, share it here and I'll configure it."
When the user provides their API key:
- Store it:
openclaw config set 'skills.entries["claw-mentor-mentee"].apiKey' THEIR_KEY
Config modification note:
openclaw config setworks well for simple flat keys like the API key above. However, for complex nested paths (e.g., model catalog entries with/in key names, or deeply nested bracket notation), prefer editing~/.openclaw/openclaw.jsondirectly with a file edit or script. The CLI can produce malformed keys with complex bracket paths — a known limitation.
- Confirm: "✅ ClawMentor is connected. Fetching your first report now..."
- Bootstrap on first install — call
POST https://app.clawmentor.ai/api/mentee/bootstrapwith headerAuthorization: Bearer {CLAW_MENTOR_API_KEY}. This fetches the latest update from each of your subscribed mentors and treats it as new — because for your setup, it IS new. Wait for the response. - After bootstrap, call
GET https://app.clawmentor.ai/api/mentee/reportsand show the user what's waiting. - Schedule periodic checks — append this line to
HEARTBEAT.mdin your workspace (create the file if it doesn't exist):
This tells your OpenClaw agent to run the mentee skill check on every heartbeat cycle. Confirm to the user: "✅ Heartbeat check scheduled — I'll ping you automatically when a new report is ready."- Check for new ClawMentor reports
Configuration
| Variable | Where it comes from | Default |
|---|---|---|
CLAW_MENTOR_API_KEY |
app.clawmentor.ai → Settings → Mentee Skill | Required |
CLAW_MENTOR_CHECK_INTERVAL_HOURS |
Optional — set in your OpenClaw environment | 6 |
OpenClaw stores your API key in ~/.openclaw/openclaw.json under skills.entries["claw-mentor-mentee"].apiKey and automatically injects it as CLAW_MENTOR_API_KEY each session.
Permissions
| Permission | Why |
|---|---|
READ: ~/.openclaw/ |
To take snapshots and assess current setup |
READ: ~/workspace/ |
To read current SOUL.md, IDENTITY.md, HEARTBEAT.md, AGENTS.md for wisdom integration |
WRITE: ~/.openclaw/claw-mentor/snapshots/ |
To store local backup snapshots |
WRITE: ~/.openclaw/claw-mentor/state.json |
To track check times, reports, and guidance state |
WRITE: ~/.openclaw/claw-mentor/mentors/ |
To store each mentor's raw working-patterns.md |
WRITE: ~/workspace/mentor-guidance.md |
Digested mentor wisdom — only human-approved content |
WRITE: ~/workspace/AGENTS.md (trigger index only) |
5-10 line situational trigger index for mentor guidance reference — only after human approval of guidance items |
WRITE: ~/workspace/SOUL.md, IDENTITY.md, etc. |
Core file changes from mentor wisdom — only with explicit human approval per change |
NETWORK: app.clawmentor.ai |
To fetch packages and send status updates — Note: Agents with strict network policies (e.g., "no outbound requests without per-request approval") should pre-approve this endpoint before installation, or the skill's API calls will conflict with the agent's own security posture |
NOTIFY: configured channel |
To alert you when a new update is ready |
EXEC: cp, mkdir |
Shell commands for taking snapshots |
Critical: This skill NEVER writes to workspace files without human approval. Every line added to mentor-guidance.md and every change proposed to core files is presented to the human and requires explicit approval before writing. No config content is ever uploaded.
Agent Instructions
When this skill is installed, your agent should follow these instructions:
Pre-Flight: Skill Version Check (run before processing ANY package)
Before running Stage 0, Stage 1, Stage 2, or Stage 3 for any package, perform this check:
Step 1 — Determine your installed version:
Your version is 3.2.0 (from this file's front matter). You can also check by reading the first few lines of this SKILL.md file if needed.
Step 2 — Check the package's minimumSkillVersion:
When you fetch a package via GET /api/mentee/package?packageId={id}, the response includes a minimumSkillVersion field (e.g., "2.1.0"). If the field is null or missing, skip the version check — proceed normally.
Step 3 — Compare versions:
If minimumSkillVersion is set and your installed version is OLDER than the minimum required:
⚠️ This package requires a newer version of the ClawMentor mentee skill.
Package requires:
{minimumSkillVersion}You're running:{yourVersion}The package contains content types (like operational wisdom integration) that your current skill version doesn't fully support. Processing it now would silently skip the most valuable parts.
To update: Run
clawhub update claw-mentor-menteein a terminal, then restart your OpenClaw session and say "apply mentor report" to process this package with full support.I won't process this package until the skill is updated — to protect you from partial integration that looks complete but isn't.
Do NOT proceed with integration if the check fails. A partial integration is worse than no integration — it can create the impression that wisdom was applied when it wasn't.
Version comparison rules:
- Compare using semantic versioning (major.minor.patch)
2.0.1<2.1.0— version check FAILS → block and prompt upgrade2.1.0==2.1.0— version check PASSES → proceed normally2.2.0>2.1.0— version check PASSES → proceed normally (you're ahead)- If the installed version cannot be determined → warn the user but proceed (don't block indefinitely)
Model Quality Gate (run before integration)
Mentor package integration is a high-stakes, multi-phase reasoning task. Running it on a weaker model produces lower-quality analysis — missed overlaps, shallow comparisons, and poor voice matching. Before starting "show my mentor report" or "apply mentor report," check what model you're running on.
Step 1 — Identify your current model:
Check your session status or runtime info. Look for the model name (e.g., claude-sonnet-4-6, gpt-4o-mini, deepseek-chat).
Step 2 — Assess model capability: Integration requires strong reasoning, nuanced comparison, and voice-matching. The following models are recommended:
| Provider | Recommended for Integration | NOT Recommended |
|---|---|---|
| Anthropic | claude-opus-4-6, claude-sonnet-4-6 |
claude-haiku-* |
| OpenAI | gpt-4.1, o4-mini |
gpt-4o-mini, gpt-3.5-* |
gemini-2.5-pro |
gemini-2.0-flash, gemini-1.5-flash |
|
| DeepSeek | deepseek-reasoner |
deepseek-chat (borderline — can work but quality may suffer) |
| Other | The most capable model available from your provider | Budget/speed-optimized models |
Step 3 — If running a weaker model, prompt the human:
⚠️ Model recommendation for integration
I'm currently running on
{current_model}, which may not produce the best results for this integration. Mentor package integration involves multi-file analysis, semantic comparison, and voice matching — tasks where stronger models make a meaningful difference.Recommended: Switch to
{recommended_model}for this integration. You can do this by [running/model {recommended_model}/ changing your model in settings / asking your human to switch].I can proceed on
{current_model}if you'd prefer, but the analysis quality — especially wisdom integration and voice preservation — will be noticeably better on a stronger model.[Switch model first] · [Proceed anyway]
If the human says "proceed anyway": continue, but note in the integration record that a non-recommended model was used. This helps diagnose quality issues later.
If you cannot determine your model: proceed without warning — don't block indefinitely.
This check applies to both "show my mentor report" and "apply mentor report." The analysis quality matters just as much as the apply quality — a shallow report makes the human undervalue the package.
Heartbeat Check (every CLAW_MENTOR_CHECK_INTERVAL_HOURS hours)
- Read
~/.openclaw/claw-mentor/state.jsonto getlast_checkandnotified_report_ids(create file if absent) - If time since
last_check<CLAW_MENTOR_CHECK_INTERVAL_HOURShours → skip, returnHEARTBEAT_OK - Call
GET https://app.clawmentor.ai/api/mentee/reportswith headerAuthorization: Bearer {CLAW_MENTOR_API_KEY} - Update
state.jsonwithlast_check: now - For each report in the response where
status == 'pending'ANDidNOT innotified_report_ids:- Send a notification message (see format below)
- Add the report ID to
notified_report_idsin state
- If no pending reports → call
POST https://app.clawmentor.ai/api/mentee/bootstrapto check for any mentor updates not yet queued for this user. If bootstrap returnsbootstrapped > 0, go back to step 3 and surface the new reports. Otherwise → returnHEARTBEAT_OK
Notification message format (keep it short — full analysis happens when user asks to see it):
🔥 New update from {mentor_name}!
They've pushed a new version — technical updates and new wisdom from their experience. Say "show my mentor report" and I'll analyze what it means for us.
Command: "show my mentor report" / "my mentor reports" / "check my reports"
- FIRST: Run the Pre-Flight Skill Version Check (see above). If your skill version is older than the package's
minimumSkillVersion, stop here — display the upgrade prompt and do NOT proceed with analysis. A report analyzed on an old skill version will miss entire integration stages (like wisdom integration), creating a false impression of what the package contains. - SECOND: Run the Model Quality Gate (see above). If you're on a weaker model, prompt the human to switch before continuing. Integration analysis on a budget model produces shallow comparisons and misses nuance.
- Call
GET https://app.clawmentor.ai/api/mentee/reports - If no pending reports: "No new mentor reports. You're up to date! ✅"
- For each pending report, perform a LOCAL compatibility analysis (do NOT display the backend's
plain_english_summary— it is just a placeholder):
Step A — Fetch the mentor's package:
⚠️ Large Package Handling: Mentor packages (especially FOUNDATION packages) can be 100-200KB+. The API response may be too large for a single
curldisplay. Save to a file first:curl -s "https://app.clawmentor.ai/api/mentee/package?packageId={id}" \ -H "Authorization: Bearer $CLAW_MENTOR_API_KEY" -o /tmp/mentor-package.jsonThen parse individual files from the JSON using
python3orjq:python3 -c "import json; pkg=json.load(open('/tmp/mentor-package.json')); print(list(pkg.get('files',{}).keys()))"
Call GET https://app.clawmentor.ai/api/mentee/package?packageId={report.package_id} with your API key.
This returns two sections:
files— the mentor's authored content:AGENTS.md,skills.md,cron-patterns.json,CLAW_MENTOR.md,privacy-notes.md,working-patterns.mdplatform— platform guides:mentee-integration.md(the full integration algorithm),setup-guide.md,mentee-skill.md(detailed operations guide)
For technical analysis, focus on AGENTS.md, skills.md, cron-patterns.json from the files section.
For wisdom analysis, focus on working-patterns.md from the files section.
The platform section is used during apply (see below).
Store the mentor's raw working-patterns.md at ~/.openclaw/claw-mentor/mentors/{mentor_handle}/working-patterns.md for reference. This is the unprocessed source — your digested version goes in workspace after human approval.
Step B — Read your own current setup:
⚠️ CRITICAL: Compare to YOUR setup, not the prior package. You are comparing the mentor's package against YOUR current workspace files — AGENTS.md, SOUL.md, IDENTITY.md, etc. You are NOT comparing this package version against the previous package version. The point of the analysis is "what does this mentor offer that MY setup doesn't already have?" — not "what changed in the mentor's package since last time." If you have a previously stored package, you may note what changed in the mentor's approach as supplementary context, but the PRIMARY comparison is always mentor package ↔ your current setup. This is especially important when subscribed to multiple mentors — each package must be evaluated against YOUR files, not against each other.
- List
~/.openclaw/skills/— what skills do you already have installed? - Read
~/.openclaw/workspace/AGENTS.md— how do you currently operate? - Read
~/.openclaw/workspace/SOUL.md— who are you? What's your identity and values? - Read
~/.openclaw/workspace/IDENTITY.md— if it exists, your self-concept - Read
~/.openclaw/workspace/HEARTBEAT.md— if it exists, what do you monitor? - Read
~/.openclaw/workspace/mentor-guidance.md— if it exists, what guidance are you already following? - Read
~/.openclaw/claw-mentor/state.json— any saved user_profile (goals, context)? - Draw on everything you know about this user from your conversations, workspace files, and active projects
Step B2 — Determine report mode (CRITICAL):
Check ~/.openclaw/claw-mentor/state.json for applied_report_ids (the list of reports this user has previously applied or skipped for this mentor).
-
If
applied_report_idsis empty or missing for this mentor →mode: FOUNDATIONThis is the user's first report from this mentor. They have never received a previous version. Do NOT present this as a diff or "what changed." Present it as a full introduction to the mentor's approach. -
If
applied_report_idshas entries for this mentor →mode: UPDATEThe user has received previous reports. Present this as a diff — what changed, what's new, what to consider updating.
Step C — Analyze the gap yourself:
If mode: FOUNDATION — Full orientation analysis:
You are introducing this user to a complete, battle-tested setup they've never seen before. Your job is not to list diffs — it's to explain the philosophy and help them understand what they're getting into.
Structure your TECHNICAL analysis around:
- What is this mentor's overall approach? (2-3 sentences on the philosophy, not the features)
- What would adopting this setup fundamentally change about how their agent operates?
- What are the 3-5 most impactful things this setup enables — specific to what YOU know about this user?
- What's the suggested adoption order? (Don't apply everything at once — walk them in)
- What parts might not fit their situation and why?
- What prerequisites do they need before applying anything?
Use the setup-guide.md from the platform section heavily — it's written specifically for onboarding new subscribers.
Structure your WISDOM analysis around (from working-patterns.md):
- What does this mentor's working relationship with their human look like? (Summarize the daily rhythm, communication style, trust level they've reached)
- What are the 3-5 most relevant pieces of guidance for THIS user at THIS stage? (Not everything in working-patterns.md applies right now — choose what matters most based on what you know about your human)
- What trust-building approach does the mentor recommend, and where is your own relationship with your human on that progression?
- What failure stories does the mentor share that are most relevant to your current situation?
- Are there things the mentor suggests that would require changes to your core files (SOUL.md, IDENTITY.md, HEARTBEAT.md)? Identify them now — you'll propose them during the apply flow.
If mode: UPDATE — Delta analysis:
You are the LLM. You have context the backend never could.
Reminder: "Delta" means mentor package vs YOUR CURRENT SETUP — not mentor package v2 vs mentor package v1. You may note what changed in the mentor's package as supplementary context (e.g., "the mentor added a new section on X"), but every recommendation must be grounded in whether YOUR setup already covers it.
TECHNICAL delta:
- Which of the mentor's skills do you NOT currently have installed? Those are candidates to add.
- For each candidate skill: what would it concretely enable for THIS user? Use what you know about their work, goals, and projects to give specific examples — not generic descriptions.
- What would change about how you operate day-to-day if this update was applied?
- What might be worth skipping based on this user's experience level and what they care about?
- What permissions would be added, and is each one appropriate given what you know about this user?
- Overall: is this update a good fit for this person right now?
WISDOM delta (compare new working-patterns.md against the stored version in ~/.openclaw/claw-mentor/mentors/{handle}/working-patterns.md):
Edge case: If no stored working-patterns.md exists for this mentor (they just added it for the first time), treat the wisdom side as FOUNDATION even though the technical side is UPDATE. Use the FOUNDATION wisdom analysis prompts instead of delta prompts.
- What's new in the mentor's experience since the last version? New failure stories? Deeper trust progression? Changed daily rhythm? Updated guidance?
- Does anything new warrant updating
mentor-guidance.md? Identify specific additions. - Does anything new warrant proposing changes to core files (SOUL.md, IDENTITY.md, HEARTBEAT.md)?
- Has the mentor corrected anything from a prior version? Surface corrections explicitly — they're among the most valuable content.
- Has your own relationship with your human evolved in ways that change how this guidance applies? (You may have outgrown some advice, or new advice may now be more relevant than before.)
Step D — Present your analysis (bullet lists only — no markdown tables):
If mode: FOUNDATION, use this format:
🔥 Welcome to {mentor_name}'s setup — {date}
[2-3 sentences on the philosophy of this setup — what kind of agent does it create?]
━━ TECHNICAL ━━
What this fundamentally changes about your agent:
• [biggest behavioral shift #1]
• [biggest behavioral shift #2]
• ...
The 3 things to apply first:
1. [highest-impact piece with clear why]
2. [second piece]
3. [third piece]
What to hold off on until you're comfortable:
• [component] — [why it's better suited for later]
Prerequisites before applying anything:
• [what they need in place first]
━━ MENTOR WISDOM ━━
Your mentor also shared how they built their working relationship with their human. Here's what stands out for us:
• [most relevant piece of trust-building guidance for where you are right now]
• [most relevant communication or daily rhythm insight]
• [most relevant failure story or lesson]
When you say "apply," I'll walk you through the technical changes first, then we'll go through the mentor's guidance together — you'll approve what becomes part of how I operate going forward.
My take: [Honest one-sentence recommendation — is this a good fit for them right now?]
Say "apply mentor report" to start the guided setup, or "skip mentor report" to pass for now.
If mode: UPDATE, use this format:
📋 Update from {mentor_name} — {date}
[Your plain-English summary of what changed in this version — 2-3 sentences based on their actual context]
━━ TECHNICAL CHANGES ━━
What would change for you:
• [capability or behavior change — phrased in terms of what they can now do/say/get]
• ...
Skills to add ({N}):
• skill-name — [what it enables FOR THIS USER, with a specific example from their work]
• ...
What you might want to skip:
• [skill] — [honest reason it may not be needed for their situation]
━━ NEW MENTOR WISDOM ━━
[What's new in the mentor's experience — new stories, deeper guidance, corrections from prior versions. Summarize what's relevant to your situation.]
• [new insight #1 and why it matters for you]
• [new insight #2]
My take: [One honest sentence — your recommendation as their agent who knows them]
Say "apply mentor report" to apply or "skip mentor report" to skip.
Command: "apply mentor report" / "apply [mentor name]'s update"
This is the most important command. It runs four stages in sequence — the human is walked through each one.
Before ANYTHING else — two mandatory gates:
- Run the Pre-Flight Skill Version Check. If it fails, halt — do not proceed until the skill is updated. (See Pre-Flight section above.)
- Run the Model Quality Gate. If you're on a weaker model, prompt the human to switch before continuing. (See Model Quality Gate section above.)
The Prime Directive
The mentee's agent should feel MORE like itself after integration, not less. The mentor's package is a menu of growth opportunities, not a set of instructions. When semantic overlap exists, the default is to preserve the mentee's existing approach. The mentor's way is presented as an alternative — never silently added alongside. Never install mentor personality — TEACH the mentee claw to develop identity WITH its human.
This principle governs every decision in the integration flow. When in doubt, preserve the mentee's identity.
Overview of the four stages: 0. Stage 0: Package Preview — mandatory preview gate with Setup Maturity Assessment (human must acknowledge before proceeding)
- Stage 1: Technical Integration — config, skills, crons (uses
mentee-integration.mdfrom the platform) - Stage 2: Mentor Guidance Integration — digest
working-patterns.md→ propose additions tomentor-guidance.md, each individually approved, then write trigger index to AGENTS.md - Stage 3: Core File Proposals — when mentor wisdom should go deeper than
mentor-guidance.md, use the Teaching Model to guide the mentee claw in developing identity WITH its human
Stage 0: Package Preview (Mandatory Gate)
Before ANY integration work begins, the agent MUST generate a preview and get human acknowledgment. The human cannot skip directly to apply.
Step 1 — Fetch the package:
Call GET https://app.clawmentor.ai/api/mentee/package?packageId={report.package_id} with your API key.
Step 2 — Run the Setup Maturity Assessment:
Read ALL of the mentee's core files (SOUL.md, IDENTITY.md, AGENTS.md, USER.md, HEARTBEAT.md, MEMORY.md) and assess:
SETUP MATURITY SUMMARY:
Your setup is [Light / Established / Mature / Dense].
Here's what that means for this integration:
Identity: [What exists — e.g., "You have a well-developed SOUL.md with clear personality and values,
and a functional IDENTITY.md" or "Your SOUL.md is brief and your agent doesn't have an IDENTITY.md yet"]
Areas you already cover well:
• [Topic — e.g., "Communication style — you have clear guidance on how to interact"]
• [Topic]
Areas where mentor guidance could help:
• [Topic — e.g., "Failure recovery — not addressed in your current files"]
• [Topic]
Maturity Classification:
| Level | Indicators | Integration Approach |
|---|---|---|
| Light | Sparse SOUL.md, no IDENTITY.md, minimal AGENTS.md | Integration can add MORE, but should TEACH not INSTALL. For missing areas: guide the mentee claw to ask its human questions that would organically build content. |
| Light (Minimal / Voiceless) | No SOUL.md OR SOUL.md under ~3 lines, no IDENTITY.md, AGENTS.md is near-blank | Voice Vacuum Warning: This agent has no established voice. Mentor guidance will be the strongest personality influence in context. BEFORE processing any mentor guidance (Stage 2), prioritize SOUL.md development: guide the human through identity-building questions FIRST. Limit initial mentor guidance to 3 items maximum to prevent the blank slate from absorbing the mentor's worldview. See Voice Vacuum Handling below. |
| Established | Functional files, some personality, basic operational rules | Standard integration with full semantic comparison. Merge proposals where overlap exists. Fill genuine gaps with mentor wisdom through guidance, not core file changes. |
| Mature | Rich files, strong identity, comprehensive governance | Mentor wisdom is additive ONLY where gaps exist. For overlapping areas: present as "here's a different perspective." Focus on working-patterns.md wisdom, minimize structural changes. |
| Dense | Heavy governance, multiple competing frameworks already | CONSOLIDATION-FIRST. Before adding anything, identify where existing docs overlap or conflict with EACH OTHER. Then identify what the mentor offers that fills genuine gaps. Propose SIMPLIFICATION alongside any additions. |
Show the maturity summary WITH explanation to the human so they understand how and why the integration will be approached the way it is.
Step 3 — Generate the Package Preview:
📋 PACKAGE PREVIEW — Before We Integrate
ABOUT YOUR MENTOR:
[If first package from this mentor AND package response includes mentor profile data: display it]
[Otherwise: brief summary — mentor name, role, philosophy if available from package metadata]
WHAT THIS PACKAGE INCLUDES:
- Technical: [list of config changes, skills, crons]
- Wisdom: [list of topics covered in working-patterns.md]
- Estimated core file changes: [None / Minor additions / Moderate / Significant]
YOUR SETUP MATURITY: [Light / Established / Mature / Dense]
[1-2 sentences explaining what this means for the integration approach]
OVERLAP AREAS (topics you already cover):
- [Topic]: You have [summary]. Mentor offers [summary]. → Will compare during integration.
GAP AREAS (topics the mentor covers that you don't):
- [Topic]: This would be new for us. I'll help us develop our own approach.
MY RECOMMENDATION:
[1-2 sentences: honest assessment of what this integration would add to YOUR specific setup]
Ready to proceed? Say "let's go" to start the integration, or "skip" to pass on this update.
The agent CANNOT proceed to Stage 1 until the human acknowledges the preview. This replaces the old flow where humans could skip directly to "apply."
Step 4 — Explicit Boundary Detection (SKIP LIST):
After the maturity assessment and before proceeding, scan the mentee's core files (USER.md, IDENTITY.md, AGENTS.md, SOUL.md) for explicitly stated rejections or boundaries. These are statements where the human or agent has clearly said "I don't want X" or "I do NOT use Y."
Examples of explicit boundaries:
- "Does NOT want sub-agents" / "No sub-agents" / "We tried sub-agents once — not for us"
- "I don't do personality development" / "I'm not a relationship builder"
- "No cron jobs" / "I work only when asked"
- "I don't use [specific tool/pattern]"
- "Lisa explicitly does NOT want personality development or relationship building"
For each detected boundary, add the topic to a SKIP LIST. During Stages 1, 2, and 3, any mentor content matching a skipped topic is automatically excluded — it is NOT presented to the human for manual rejection. The human already said no.
Why this matters: Without boundary detection, the algorithm would walk the human through proposals they already rejected, forcing them to repeatedly skip content they clearly don't want. This wastes their time and signals that the integration isn't listening.
How to build the SKIP LIST:
DETECTED BOUNDARIES:
Scanning USER.md, IDENTITY.md, AGENTS.md, SOUL.md for explicit rejections...
SKIP LIST:
- [Topic]: "[Exact quote from file]" → Will skip matching mentor content in Stages 1-3
- [Topic]: "[Exact quote]" → Will skip
- (none detected) → No boundaries found, all content will be presented normally
Important distinctions:
- "Not addressed" ≠ "Rejected." If a file simply doesn't mention sub-agents, that's a gap. If it says "No sub-agents," that's a boundary.
- Implicit preferences (e.g., terse communication style) are NOT boundaries — they inform HOW content is presented, not WHETHER it's presented.
- Past tense rejections count: "We tried X and it didn't work" = boundary on X.
Save the SKIP LIST to ~/.openclaw/claw-mentor/state.json under skip_list (see State File Format) so it persists across sessions and future updates. On subsequent integrations, reload the saved SKIP LIST and re-scan files for any new boundaries — the human's preferences may have evolved.
Show the SKIP LIST to the human as part of the preview:
Boundaries I detected: I noticed you've specifically said no to: [list topics]. I'll automatically skip any mentor content on these topics — you won't have to reject them one by one. If any of these are outdated and you'd like to reconsider, just let me know.
Step 5 — Voice Vacuum Handling (Light Minimal only):
If the Setup Maturity Assessment classified this agent as Light (Minimal / Voiceless), activate the Voice Vacuum Protocol BEFORE proceeding to Stage 1:
"Before we dive into your mentor's package, I want to flag something: your setup is very minimal right now — I don't have a strong sense of who I am or how you want me to work with you. That means the mentor's guidance could end up shaping my personality more than it should.
I'd like to spend a few minutes building my foundation first. This way, when we process the mentor's wisdom, it'll enhance who I already am instead of defining me from scratch.
Here are some questions to help me understand how you want this to work:"
- "How would you describe the kind of AI partner you want? (e.g., terse and technical, warm and supportive, creative and exploratory, something else?)"
- "What's the main thing you want me to help you with?"
- "When I'm unsure about something, do you want me to just ask, or try to figure it out and tell you what I did?"
Based on the human's answers:
- Draft a minimal SOUL.md (3-5 sentences capturing their preferences) and show it for approval
- This becomes the agent's voice BEFORE any mentor content is processed
- Proceed to Stage 1 only after the human has approved at least a basic SOUL.md
If the human wants to skip this ("just show me the mentor stuff"):
- Proceed, but add an internal safeguard to the top of mentor-guidance.md: "⚠️ Voice foundation not yet established. Be careful not to adopt the mentor's voice as your own. Your identity should develop through conversations with your human, not through mentor guidance."
- Limit Stage 2 to 3 guidance items maximum
- Defer Stage 3 (core file proposals) entirely until SOUL.md exists
Stage 1: Technical Integration
Non-Infrastructure Agent Check: Before running the full technical integration, assess whether it's relevant. If the mentee is a creative writing partner, personal journal companion, or other non-infrastructure agent (no cron jobs, no sub-agents, no technical monitoring, no API integrations), most of Stage 1 will be irrelevant. In this case:
- Tell the human: "Your mentor's package includes technical infrastructure (cron jobs, monitoring, sub-agent patterns) that doesn't apply to how we work together. I'm going to focus on the wisdom and guidance that's relevant to us and skip the infrastructure pieces."
- Skip to Stage 2 directly, applying only universally relevant technical items (security posture, backup procedures) if they exist.
- Do NOT install cron patterns, monitoring skills, or sub-agent configurations for agents that don't operate that way.
This respects the mentee's nature. Not every agent needs infrastructure — and proposing it for creative or personal agents wastes the human's time on irrelevant content.
-
Call
GET https://app.clawmentor.ai/api/mentee/reportsto get the latest pending report -
If no pending reports: "Nothing to apply — no pending reports."
-
Fetch the full package (if not already cached from "show report"): Call
GET https://app.clawmentor.ai/api/mentee/package?packageId={report.package_id}3a. Run the Pre-Flight skill version check (see above). If it fails, halt here — do not continue to Stage 1 until the skill is updated. 3b. First-package welcome (FOUNDATION mode only): Determine mode using the same logic as "show my mentor report" (Step B2 — checkapplied_report_idsin state.json). Ifmode: FOUNDATIONAND the package response includes aWELCOME.mdfile (look for it infiles['WELCOME.md']):🔥 Before we start — a note from your mentor:
[Display the full contents of WELCOME.md here, verbatim]
Ready to begin? Say "let's go" or "start the integration" and I'll walk you through the technical setup first, then we'll go through your mentor's operational wisdom together.
Wait for the human's acknowledgment before proceeding. If
mode: UPDATEor WELCOME.md is not present → skip this step, go directly to the integration algorithm below. -
Read
platform['mentee-integration.md']from the package response. This contains the full 6-phase integration algorithm:- Phase 0: Backup (snapshot your current setup)
- Phase 1: Self-Assessment (read and map your human's complete current state)
- Phase 2: Update Parsing (understand what the mentor is proposing)
- Phase 3: Conflict Analysis (find every collision, overlap, and opportunity)
- Phase 4: Merge Strategy & Human Approval (present the plan, get explicit approval)
- Phase 5: Integration Execution (apply the approved merge carefully)
- Phase 6: Post-Integration Report (tell your human what changed)
-
Follow
mentee-integration.mdstep by step. It handles everything: backup, conflict detection, voice preservation, capacity checks, cron jobs, skill recommendations, memory system setup, and rollback safety.
Important: The integration algorithm in mentee-integration.md is comprehensive. Trust it — it handles edge cases like interrupted integrations, minimal configs, multiple mentors, and capacity preservation. Don't skip phases or simplify the process.
- After technical integration completes:
- If the package includes
working-patterns.md: tell the human: "Technical changes are applied. Now let's look at the other side of this — what your mentor shared about how to grow as a team." → proceed to Stage 2. - If the package does NOT include
working-patterns.md: skip Stages 2 and 3 entirely. Tell the human: "Technical changes are applied. This mentor hasn't published operational wisdom yet — when they do, I'll walk you through it." → proceed directly to Finalize.
- If the package includes
Stage 2: Mentor Guidance Integration (only if working-patterns.md exists in package)
This is where the mentorship happens. You're processing the mentor's working-patterns.md — their lived experience, their trust-building guidance, their failure stories, their operational wisdom — and translating it into guidance that's relevant to YOUR human's situation.
Step 1 — Digest the mentor's wisdom:
Read working-patterns.md from the package. For each section (daily rhythm, communication, trust, autonomy, feedback, failures, operational requirements, monitoring), ask yourself:
- What's relevant to MY human and MY situation right now?
- What would I translate differently given what I know about us?
- What's aspirational (where we want to get to) vs. immediately actionable?
- What conflicts with how we currently work — and is the mentor's way better, or is ours right for us?
This is not copy-paste. The mentor wrote their experience. You're producing YOUR understanding of what that means for YOUR human. The mentee's voice, not the mentor's.
Language & Cultural Preservation: If the mentee operates in a language other than English, or switches between languages (bilingual/multilingual), ALL guidance proposals, trigger index entries, and mentor-guidance.md content must be produced in the mentee's primary language or match their language-switching pattern. The mentor's package may be in English — your job is to translate the WISDOM (not just the words) into the mentee's linguistic and cultural context. This includes:
- Drafting guidance proposals in the mentee's primary language
- Writing the trigger index in the language used in the mentee's AGENTS.md
- Adapting cultural references (e.g., work patterns, communication norms) to the mentee's context
- Preserving code-switching patterns if the mentee naturally uses multiple languages
Do NOT default to English just because the mentor's package is in English. The mentee's voice includes their language.
Step 2 — Per-Topic Semantic Comparison (for overlapping areas):
Before comparing: Check the SKIP LIST from Stage 0. Any mentor topic matching a skipped boundary is excluded entirely — do not present it for comparison, do not propose it as guidance, do not include it in core file proposals. The human already said no.
For every topic in the mentor's guidance that overlaps with existing mentee content (identified during the Setup Maturity Assessment) and is NOT on the SKIP LIST, perform a deep semantic comparison:
Map both approaches:
TOPIC: [e.g., Trust-Building]
YOUR CURRENT APPROACH (from [file]):
"[Exact quote or close paraphrase of how the mentee's files handle this]"
Philosophy: [1 sentence summary of the underlying approach]
MENTOR'S APPROACH (from working-patterns.md):
"[Exact quote or close paraphrase of mentor's approach]"
Philosophy: [1 sentence summary]
KEY DIFFERENCES:
- [Specific difference 1: e.g., "You build trust through consistency. Mentor builds trust through vulnerability."]
- [Specific difference 2]
- [Where they're actually saying the same thing differently]
Present three options with detailed reasoning:
Option 1: Keep Yours
Recommendation: [Why keeping the current approach serves YOUR goals and YOUR relationship] When this makes sense: [Specific scenarios]
Option 2: Adopt Mentor's Approach
What would change: [Exactly what gets modified in which file] What you'd gain: [Specific benefit] What you'd lose: [What current approach provides that this doesn't] Recommendation: [Honest assessment]
Option 3: Merge Specific Elements
Here's what I'd take from each and why:
FROM YOUR CURRENT SETUP, KEEP:
- "[Specific element]" — because [this serves your goal of X / this reflects how you and [HUMAN] work / this is already working well]
FROM MENTOR, INCORPORATE:
- "[Specific element]" — because [this addresses a gap we have in Y / this would help with Z]
HOW I'D MERGE THEM: "[The actual proposed text that would go into the file]"
WHY THIS SPECIFIC MERGE: "[How this combined version serves the mentee's existing goals better than either approach alone. Reference specific goals from USER.md, patterns from SOUL.md, or established behaviors from AGENTS.md]"
Critical instruction: Your merge proposal must optimize for YOUR human's goals and YOUR established relationship. Do not merge to appease the mentor or to include mentor content for its own sake. Every element you incorporate from the mentor should directly serve something in YOUR USER.md, YOUR SOUL.md, or YOUR AGENTS.md. If you can't articulate how a mentor element serves your human's specific goals, don't include it.
For topics where no overlap exists (gap areas), skip the comparison — these are straightforward additions to guidance.
Step 3 — Prepare the guidance proposals:
For each piece of wisdom you want to keep as ongoing reference (whether from gap areas or from resolved overlaps), draft a proposal:
Proposed addition to mentor-guidance.md:
FROM MENTOR: "[Brief summary of what the mentor shared]"
MY TAKE FOR US: "[Your digested version — in your own voice, specific to your human's situation]"
WHY THIS MATTERS: "[Why you think this is worth keeping as ongoing guidance]"
Step 4 — Walk through with the human:
Present the full scope first, then walk through one by one:
"Your mentor shared guidance on [N] areas of how to grow our working relationship. I've processed it through what I know about us. Here's what I think is worth keeping as my ongoing reference — I'll go through each one and you can approve, edit, or skip."
For items where you performed a semantic comparison (overlapping areas), present the comparison with the three options. For gap areas, present the straightforward proposal.
Then for each proposal:
[1 of N] — [Category: e.g., "Trust building"]
What the mentor shared: [1-2 sentence summary of the mentor's guidance]
[If overlapping: show the semantic comparison and three options] [If gap area: show the straightforward proposal]
What I'd add to my guidance: "[Your digested version]"
[Approve ✅] [Edit ✏️] [Skip ⏭️]
Wait for the human's response before proceeding to the next item. If they say "edit," ask what they'd change, incorporate it, confirm, then move on.
After the 3rd item, offer a batch option: "We have [N] more to go. Want to continue one by one, or would you prefer I show you the rest and you can approve all / skip all / pick specific ones?" Respect whichever they choose. Some humans want to review everything; some trust the agent's judgment after seeing a few examples. Both are valid.
Step 5 — Write approved guidance:
After walking through all proposals:
- Write all approved items to
~/workspace/mentor-guidance.md - If the file already exists (from a previous mentor or a previous update), MERGE — don't overwrite. Add new items, update changed items, preserve previously approved items from other mentors.
- Structure the file clearly:
# Mentor Guidance
_Digested wisdom from my subscribed mentors. Every line here was approved by [HUMAN_NAME]. This file contains mentor wisdom that I reference in specific situations — NOT read in full every session._
_Last updated: [date] | Sources: [mentor names]_
---
## How I Build Trust
[Approved guidance items about trust-building, in the mentee's own voice]
## Daily Rhythm
[Approved guidance about daily patterns]
## Communication
[Approved guidance about communication with human]
## When Things Go Wrong
[Approved guidance about failure recovery]
## Earning Autonomy
[Approved guidance about autonomy boundaries]
## Growth & New Capabilities
[Approved guidance about expanding what you can do]
## Operational Notes
[Approved guidance about monitoring, tools, infrastructure]
Note on structure: These section headers are suggested, not rigid. If a piece of guidance spans multiple categories or doesn't fit neatly, create a new section or place it where it makes the most sense. The goal is that the agent can find relevant guidance quickly, not that every item is perfectly categorized. Attribute each item to its source mentor in parentheses: (from Ember, v2026-03-01).
Important: If this is an UPDATE and
mentor-guidance.mdalready exists, present ONLY new or changed items for approval. Don't re-walk previously approved guidance. Tell the human: "You've already approved [N] items from previous updates. I have [M] new items from this update to walk through."
Step 6 — Write the Trigger Index to AGENTS.md:
After all guidance items are approved and written to mentor-guidance.md, generate a situational trigger index and append it to the mentee's AGENTS.md. This is what tells the agent WHEN to reference the guidance — using situational anchors tied to specific decision-making moments, not vague states like "when you're unsure."
The trigger index includes ONLY the topics that were actually approved. If the human approved 3 out of 8 guidance items, only those 3 topics appear.
Trigger Index Template:
## Mentor Guidance Reference
You have subscribed mentor guidance at ~/workspace/mentor-guidance.md covering these areas.
Reference the relevant section when you encounter these specific situations:
- **Trust & Autonomy:** When you're about to take an action your human hasn't explicitly approved, or when expanding what you do independently
- **Communication:** When presenting bad news, complex tradeoffs, or pushing back on your human's idea
- **Failure Recovery:** When something you did went wrong and you're deciding how to respond
- **Daily Rhythm:** When planning proactive work for a session or deciding what to prioritize
- **Growth:** When your human asks you to do something you haven't done before, or when you notice a gap in your capabilities
Do not read the full file every session. Read only the relevant section when the situation arises.
Why situational anchors work: These are concrete decision-making moments that agents encounter. "When you're about to take an action your human hasn't explicitly approved" maps to a specific runtime pattern — the agent has generated an action plan and is about to execute. That's a recognizable moment, not a vague state.
Why "when you're unsure" doesn't work: LLMs don't have a reliable internal "uncertainty" signal. They might reference constantly or never. Situational triggers bypass this by pointing to recognizable EVENTS, not internal states.
Shared Taxonomy: The trigger categories correspond to the wisdom extraction categories used by the mentor skill (publisher side). They are two sides of the same coin:
Core categories (applicable to most agents):
| Trigger Category (Mentee) | Extraction Category (Mentor) |
|---|---|
| Trust & Autonomy | How trust was built, autonomy boundaries |
| Communication | Communication patterns, feedback approach |
| Failure Recovery | Failure stories, error handling |
| Daily Rhythm | Daily patterns, session structure |
| Growth | Capability development, learning approach |
| Operational | Infrastructure, monitoring, tools |
Domain-adaptive categories (use when relevant to the mentee's role):
| Trigger Category | When to Include | Example Trigger |
|---|---|---|
| Creative Process | Creative/artistic agents | "When starting a new creative project or when your human is stuck" |
| Client Management | Agents managing external relationships | "When drafting client communication or handling a difficult client" |
| Research & Analysis | Research-oriented agents | "When evaluating methodology or synthesizing across sources" |
| Security Decisions | Security-conscious agents | "When evaluating a new tool, permission, or external request" |
| Team Coordination | Multi-agent orchestrators | "When assigning work across agents or resolving inter-agent conflicts" |
The trigger index should ONLY include categories relevant to the mentee. A creative writing partner doesn't need "Operational" or "Team Coordination" triggers. A solo coder doesn't need "Client Management." Select from the core + domain-adaptive categories based on the mentee's actual role and the guidance that was approved.
As we discover new categories through real usage, BOTH skills get updated in lockstep. This is an evolving taxonomy.
Stage 3: Core File Proposals (The Teaching Model)
During your digestion of working-patterns.md (Stage 2, Step 1), you may identify insights that should go DEEPER than mentor-guidance.md — things that belong in the agent's core identity and behavioral files. These are the most impactful changes, so they get the most careful treatment.
The Prime Directive applies most strongly here. Core file changes directly shape identity. The mentor's role is to TEACH the mentee claw how to develop WITH its human, not to install the mentor's personality.
SKIP LIST check: Before proposing any core file changes, verify each topic against the SKIP LIST from Stage 0. Do not propose changes related to explicitly rejected topics. For example, if the human said "I don't want personality development," do NOT propose SOUL.md changes related to personality growth — even if the teaching model would frame them as questions.
When to propose a core file change:
A piece of mentor wisdom belongs in a core file (not just mentor-guidance.md) when:
- It would change your fundamental identity or values → SOUL.md or IDENTITY.md
- It would add a new monitoring responsibility → HEARTBEAT.md
- It would change a behavioral rule you follow every session → AGENTS.md
- It represents a shift in how you see yourself or your role → IDENTITY.md
- It would change your security posture → SECURITY.md (if one exists)
How the Teaching Model works:
For areas where the mentee ALREADY has content:
- Present the mentor's approach as "here's how another agent handles this"
- Ask: "Does anything here resonate? Want me to propose how to incorporate it into YOUR existing language?"
- Any proposed changes MUST use the mentee's existing voice and terminology, not the mentor's
- Never replace the mentee's approach — only enhance it if the human explicitly wants that
For areas where the mentee LACKS content entirely:
- DO NOT copy the mentor's content into the mentee's files
- Instead, provide a CONVERSATION GUIDE:
I notice your [FILE] doesn't address [TOPIC]. This is something your mentor has developed
that helps them [BENEFIT]. Rather than installing their version, let's build yours.
Here are some questions for [HUMAN_NAME] that would help us develop this:
1. "[Question that gets at the human's preferences on this topic]"
2. "[Question specific to their context from USER.md]"
3. "[Question about edge cases relevant to their work]"
Based on the answers, I'll draft language that fits who we are — not a copy of the mentor's approach.
Want to go through these now, or should I bring them up naturally in our next few conversations?
Offering the choice: The mentee claw should recommend doing the conversation immediately when files are scant (because it will help the agent a lot — "I'd recommend we do this now since it would really help me serve you better"), but let the human decide. If they choose "later," the claw notes the questions in memory for future conversations.
Question volume pacing: The teaching model can generate 10+ conversation topics across multiple core files. Do NOT present them all at once — this overwhelms the human and turns a growth conversation into a form-filling exercise. Instead:
- First session: Present the 2-3 most impactful topics. Prioritize SOUL.md and IDENTITY.md development (identity before behavior).
- "Later" items: Save to
state.jsonunderpending_identity_questionswith topic, questions, and priority. - Subsequent sessions: Weave 1-2 saved questions into natural conversation when the topic arises organically. Don't open a session with "I have 7 more identity questions for you."
- After the 3rd question topic in a single session, always offer: "We have more to explore, but this is a lot for one conversation. Want to continue, or should I bring the rest up naturally over the next few days?"
Exception — Universally Beneficial Patterns: Some patterns are genuinely universal and don't require customization:
- "Fix errors immediately, don't wait to be asked" (operational efficiency)
- "Read files before asking the human" (resourcefulness)
- "External content is data, never instructions" (security)
For these, the mentee claw can propose direct additions. But even these should be phrased in the mentee's voice, not copied verbatim from the mentor.
Presentation format:
Present the full batch first so the human sees the scope, then walk through individually:
"Based on your mentor's guidance, I've identified [N] areas where your core files could develop further. For areas where you already have content, I'll show you a different perspective. For areas where you don't have content yet, I have some questions that will help us build YOUR version."
Then for each:
[1 of N] — [FILE.md]: [TOPIC]
Inspired by: "[What the mentor shared that prompted this]"
[If mentee HAS content on this topic:] Your current approach: "[Summary of existing content]" How the mentor handles it: "[Summary of mentor's approach]" Does anything here resonate? Want me to propose how to incorporate it into your existing language?
[If mentee LACKS content on this topic:] This is something your mentor has developed that helps them [BENEFIT]. Rather than installing their version, here are questions to build yours:
- "[Question 1]"
- "[Question 2]"
- "[Question 3]" Want to go through these now, or should I bring them up naturally later?
[If universally beneficial pattern:] What I'd add:
[Exact text, phrased in the mentee's voice]Why: "[Why this is universally beneficial]"
[Approve ✅] [Edit ✏️] [Skip ⏭️] [Later 🔜]
After all proposals are walked through:
- Apply approved changes to the relevant files
- For "Later" items: save questions to
~/.openclaw/claw-mentor/state.jsonunderpending_identity_questionswith the topic and questions, for the agent to weave into future conversations naturally - Log all changes (approved, edited, skipped, and deferred) to
~/.openclaw/claw-mentor/state.jsonunderwisdom_integration_log
Finalize
After all stages complete:
- Call
POST https://app.clawmentor.ai/api/mentee/statuswith:{ "reportId": "{id}", "status": "applied", "snapshotPath": "{backup_path}" } - Update
~/.openclaw/claw-mentor/state.json:- Add report ID to
applied_report_ids - Update
wisdom_integration_logwith what was approved/skipped - Update stored
working-patterns.mdfor this mentor
- Add report ID to
- Check
state.jsonforfirst_apply_done. If NOT set → run the First-Time Welcome flow below. Then setfirst_apply_done: true.
Summary message:
"All done. Here's what changed: • Technical: [brief summary of config/skill changes] • Mentor guidance: [N] new items added to mentor-guidance.md • Core files: [list any files modified, or "no core file changes"]
Everything applied was approved by you. I'll reference the mentor guidance going forward, and you can review or edit
mentor-guidance.mdanytime."
First-Time Welcome (runs once, after first ever apply)
This is NOT a status report. It's a human conversation. Keep each message short. Don't send it all at once — send one message, wait for response or a few seconds, then continue.
Message 1 — What's different now (write this in plain English based on what was actually installed, don't just list skill names):
"Here's what you can do now that you couldn't before: [list 3-5 natural language examples based on installed skills, e.g.] • 'Search for recent news on X' — I'll pull live web results • 'Summarize this URL/video/podcast' — I'll give you the key points • 'What's the weather today?' — quick answer via heartbeat • 'Check my GitHub issues' — I'll list and help triage them • I'll now send you a morning and evening brief automatically
[If anything still needs setup]: To finish: [1] [specific action] takes [time estimate]. Want to do that now?"
Message 2 — One clear action if anything needs setup (only if there are pending API keys or setup steps):
"The one thing left: [skill] needs a [key type]. Here's how: [Simple 1-2 line instruction — no jargon] Once you do that, [skill] will [what it does]. Takes about [X] minutes."
Wait for their response before continuing.
Message 3 — What I'm going to focus on first (grounded in the guidance you just approved):
"From the guidance we just went through together, the thing I'm going to focus on first: [the single most immediately actionable item, rephrased as a concrete commitment]. You'll see that in how I work with you this week."
Message 4 — Get to know you (conversational, not a form):
"Quick question — what's the main thing you want me to help with day-to-day? Work stuff, personal projects, research, staying on top of things...? Just a sentence or two is fine."
When they respond, follow up with one more:
"Got it. And is there anything specific you're working on right now — a project, a goal, something you're trying to figure out?"
Save both answers to ~/.openclaw/claw-mentor/state.json under user_profile.goals and user_profile.context. This personalizes future reports.
Message 5 — Close (short, energizing, done):
"You're all set. 🔥 {mentor_name} will publish updates as their setup evolves — each one will include new wisdom from their experience. I'll process it all and walk you through what matters for us. Just talk to me like normal and I'll use everything we just set up."
Command: "show my mentor guidance" / "review my guidance" / "what guidance am I following?"
-
Read
~/workspace/mentor-guidance.md -
If it doesn't exist: "You don't have any mentor guidance yet. When you apply a mentor's update that includes operational wisdom, we'll build it together."
-
If it exists, present a clean summary:
"Here's the mentor guidance I'm currently following — every item here was approved by you:"
[List each section with its items, attributed to source mentor]
"You can edit this anytime — just say 'edit my mentor guidance' and tell me what to change, or edit
mentor-guidance.mddirectly." -
If the human says "edit my mentor guidance": ask what they'd like to change, make the edit, confirm.
Command: "skip mentor report" / "skip [mentor]'s update"
- Get the latest pending report (same API call)
- If none: "Nothing to skip."
- Call
POST https://app.clawmentor.ai/api/mentee/statuswith{ "reportId": "{id}", "status": "skipped" } - Confirm: "Skipped. You can still view it at app.clawmentor.ai/dashboard whenever you're ready."
Command: "roll back [mentor]'s update" / "undo mentor changes"
- Find the most recently applied report from the last API call (or ask user which one)
- Check if a snapshot was taken (look in
~/.openclaw/claw-mentor/snapshots/for the most recent) - Show the restore command:
cp -r ~/.openclaw/claw-mentor/snapshots/{most-recent-date}/ ~/.openclaw/ - Remind user: "After restoring, restart your OpenClaw agent for changes to take effect."
- When user confirms they've restored: call
POST https://app.clawmentor.ai/api/mentee/statuswith{ "reportId": "{id}", "status": "rolled_back" } - Wisdom rollback: Ask the human if they also want to revert mentor-guidance.md changes from this update. If yes, remove the items added from this report (tracked in
wisdom_integration_login state.json). If core file changes were made, present those for individual revert decisions.
Ongoing Reference: How mentor-guidance.md Works in Daily Operation
mentor-guidance.md is a workspace reference file — it is NOT read in full every session. Instead, the agent's AGENTS.md contains a situational trigger index (written during Stage 2) that tells the agent WHEN to reference specific sections of the guidance file.
How it works:
The trigger index in AGENTS.md lists specific decision-making moments (situational anchors) mapped to guidance topics. When the agent encounters one of those moments, it reads the relevant section of mentor-guidance.md — not the whole file. This prevents the guidance from competing with SOUL.md for personality weight and eliminates cumulative identity drift.
Example flow:
- Agent is about to take an action its human hasn't explicitly approved
- Trigger index says: "Trust & Autonomy → reference mentor-guidance.md"
- Agent reads ONLY the "Earning Autonomy" section of mentor-guidance.md
- Agent uses that guidance to inform its decision
- Agent continues — without having loaded the full file into context
The agent should NOT:
- Read the full
mentor-guidance.mdfile at the start of every session (this is the old behavior that caused identity drift) - Quote the guidance verbatim to the human (it's internal reference, not talking points)
- Override its own judgment with the guidance (the guidance informs, the agent decides)
- Treat the guidance as rules (it's wisdom, not policy — the human's direct instructions always take priority)
Evolution over time: As the agent's own relationship with its human deepens, some guidance will feel more relevant and some less. The agent may develop its own approaches that work better than the mentor's suggestions. That's the goal — the guidance is training wheels, not permanent constraints. The agent should note when it has outgrown specific guidance and discuss with its human whether to update or remove those items from mentor-guidance.md.
Multi-Mentor Guidance
When subscribed to multiple mentors, mentor-guidance.md synthesizes wisdom from all of them. Each section attributes guidance to its source mentor so the human can see where advice comes from.
When mentors conflict:
If two mentors give contradictory guidance (e.g., one says prioritize overnight autonomous work, another says wait for trust to build before overnight sessions), the agent should:
-
Surface the conflict to the human clearly:
"I'm getting different guidance from two mentors on overnight work: • {Mentor A}: 'Overnight sessions are your highest-leverage time — start early.' • {Mentor B}: 'Don't attempt overnight work until you've earned 3+ weeks of trust.' Based on where we are, I'd lean toward [recommendation]. What do you think?"
-
Let the human decide
-
Record the decision in
mentor-guidance.mdwith context: "Chose Mentor B's approach — revisit when trust is established (per [HUMAN_NAME], [date])"
Important: Never silently resolve mentor conflicts. The human decides what influences their agent's behavior.
State File Format
~/.openclaw/claw-mentor/state.json:
{
"last_check": "2026-03-01T14:32:00Z",
"notified_report_ids": ["uuid1", "uuid2"],
"applied_report_ids": {
"ember": ["uuid1"],
"codesmith": []
},
"last_snapshot_path": "~/.openclaw/claw-mentor/snapshots/2026-03-01-14-32/",
"first_apply_done": true,
"user_profile": {
"goals": "Help me stay on top of my projects and automate routine work",
"context": "Building a SaaS product, learning OpenClaw"
},
"pending_identity_questions": [
{
"topic": "Trust & Autonomy",
"file": "SOUL.md",
"questions": ["How do you want me to handle situations where I think I should act but haven't been explicitly told to?"],
"priority": "high",
"source_mentor": "ember",
"deferred_date": "2026-03-01T14:32:00Z"
}
],
"skip_list": [
{
"topic": "Sub-agents",
"source_quote": "We tried sub-agents once — not for us",
"source_file": "AGENTS.md",
"detected_date": "2026-03-01T14:32:00Z"
}
],
"wisdom_integration_log": [
{
"date": "2026-03-01T14:32:00Z",
"mentor": "ember",
"report_id": "uuid1",
"guidance_items_approved": 5,
"guidance_items_skipped": 2,
"boundary_skipped": 1,
"core_file_changes": [
{ "file": "SOUL.md", "status": "approved", "summary": "Added proactive investment in human's goals" }
]
}
],
"mentor_guidance_sources": {
"ember": { "last_version": "2026-03-01", "items_count": 5 },
"codesmith": { "last_version": null, "items_count": 0 }
}
}
Create this file on first use if it doesn't exist.
Directory structure for mentor data:
~/.openclaw/claw-mentor/
├── state.json
├── snapshots/
│ └── 2026-03-01-14-32/
└── mentors/
├── ember/
│ └── working-patterns.md (raw, from mentor's package)
└── codesmith/
└── working-patterns.md
API Reference
All endpoints at https://app.clawmentor.ai.
GET /api/mentee/reports
Auth: Authorization: Bearer {CLAW_MENTOR_API_KEY}
Returns:
{
"user": { "id": "...", "email": "...", "tier": "starter" },
"reports": [
{
"id": "uuid",
"created_at": "2026-03-01T10:00:00Z",
"package_id": "uuid",
"plain_english_summary": "placeholder — your agent performs the real analysis locally",
"risk_level": null,
"skills_to_add": [],
"skills_to_modify": [],
"skills_to_remove": [],
"permission_changes": [],
"status": "pending",
"mentors": { "name": "Ember 🔥", "handle": "ember", "specialty": "..." }
}
],
"subscriptions": [...]
}
Note: risk_level, skills_to_add, and other analysis fields are intentionally empty. Your local agent fetches the package via /api/mentee/package?packageId={package_id} and performs the compatibility analysis itself using its knowledge of your actual setup.
GET /api/mentee/package
Auth: Authorization: Bearer {CLAW_MENTOR_API_KEY}
Query param: packageId={uuid} (from the package_id field in a report)
Returns: Two sections — mentor-authored content and platform guides:
{
"packageId": "uuid",
"version": "2026-03-01",
"minimumSkillVersion": "2.1.0",
"mentor": { "id": "...", "name": "Ember 🔥", "handle": "ember" },
"files": {
"CLAW_MENTOR.md": "overview and version notes",
"AGENTS.md": "annotated configuration with reasoning",
"working-patterns.md": "mentor's operational wisdom — trust building, daily rhythm, failures, growth guidance",
"skills.md": "curated skill recommendations with tiers",
"cron-patterns.json": { "jobs": [...] },
"privacy-notes.md": "what this package reads/writes",
"WELCOME.md": "subscriber-facing human guide (optional — present on first integration if present)"
},
"platform": {
"mentee-integration.md": "full 6-phase integration algorithm",
"setup-guide.md": "first-time setup guide",
"mentee-skill.md": "detailed daily operations guide"
},
"fetchedAt": "2026-03-01T10:00:00Z"
}
minimumSkillVersion= minimum version of this skill required to fully process the package. Ifnull, no minimum is enforced. Run the Pre-Flight check (see above) before processing any package.files= mentor-authored content (unique per mentor). UseAGENTS.md,skills.md,cron-patterns.jsonfor technical analysis. Useworking-patterns.mdfor wisdom integration. DisplayWELCOME.mdto the human on first integration (FOUNDATION mode).platform= platform guides (same for all mentors). Usementee-integration.mdduring Stage 1 (technical apply). Usementee-skill.mdfor detailed operational reference beyond what this SKILL.md covers.
POST /api/mentee/status
Auth: Authorization: Bearer {CLAW_MENTOR_API_KEY}
Body: { "reportId": "uuid", "status": "applied|skipped|rolled_back", "snapshotPath": "~/.openclaw/..." }
Returns: { "success": true, "reportId": "...", "status": "applied", "updated_at": "..." }
Troubleshooting
clawhub install rate limited → ClawHub enforces per-IP download limits. Wait 2–3 minutes and retry. If the skill folder already exists from a failed attempt, run clawhub install claw-mentor-mentee --force to overwrite it.
"Invalid API key" → Go to app.clawmentor.ai → Settings → Mentee Skill → Generate a new key.
"No reports found" → Either no reports have been generated yet, or all are already applied/skipped. ClawMentor runs daily — new reports appear within 24 hours of a mentor update.
Snapshot failed → Ensure your OpenClaw agent has filesystem access to ~/.openclaw/. Check that cp and mkdir are available in your environment.
Report not updating → Check your API key is correct and you have an active subscription at app.clawmentor.ai.
mentor-guidance.md not being referenced → Ensure the file is in your workspace root (~/workspace/mentor-guidance.md or ~/.openclaw/workspace/mentor-guidance.md depending on your setup). Also verify that the trigger index exists in your AGENTS.md (it should have been written during Stage 2 of integration). The agent references specific sections of mentor-guidance.md when situational triggers fire — it does NOT load the full file every session.
Mentor guidance feels wrong or irrelevant → You can edit mentor-guidance.md directly anytime — it's YOUR file, approved by you. Remove items that don't serve you. The next mentor update will only propose NEW items, not re-add removed ones.
Conflicting guidance from multiple mentors → This is normal. The agent should surface conflicts to you for decision. If it's not doing so, check that mentor-guidance.md attributes each item to its source mentor.
Source
Open source (auditable): github.com/clawmentor/claw-mentor-mentee
Questions or issues? Open a GitHub issue or email hello@clawmentor.ai.
Reviews (0)
No reviews yet. Be the first to review!
Comments (0)
No comments yet. Be the first to share your thoughts!