🧪 Skills

Github Trend Observer

A GitHub Intelligence Tool from an AI PM perspective. Goes beyond displaying data to deliver PM-grade paradigm insights. Powered by the locally authenticated...

v0.1.1
❤️ 0
⬇️ 1
👁 1
Share

Description


name: GitHub Radar description: > A GitHub Intelligence Tool from an AI PM perspective. Goes beyond displaying data to deliver PM-grade paradigm insights. Powered by the locally authenticated gh CLI + GitHub API. version: 0.1.1 author: Kun tags: [github, intelligence, pm-insight, trending, ecosystem-analysis] categories: [research, developer-tools, product-intelligence] tools:

  • name: gh version: ">=2.40.0" type: cli required: true install_url: https://cli.github.com auth_required: true check: "gh auth status"
  • name: python version: ">=3.9" type: runtime required: true check: "python --version" packages: [] # stdlib only, no pip dependencies

GitHub Radar

An open-source intelligence engine for AI PMs. Four modes, one Layer analysis framework.

Language Selection

This skill ships with both English and Chinese versions. The agent should automatically match the user's language:

  • If the user speaks English -> Use skill.md, agents/analyzer.md, references/layer_model.md, templates/*.html, and generate English reports
  • If the user speaks Chinese -> Use skill_cn.md, agents/analyzer_cn.md, references/layer_model_cn.md, templates/*_cn.html, and generate Chinese reports
  • Scripts (scripts/) and config (config/) are language-neutral and shared by both versions

When to Use

  • "What's worth looking at today?" / --pulse -> Mode 1
  • "Find me GitHub projects related to [topic]" -> Mode 2
  • "Monitor anomalous signals" / --watch -> Mode 3
  • "Analyze the ecosystem around [repo]" -> Mode 4
  • Any need involving GitHub project discovery, trend analysis, or paradigm assessment

File Structure

github-trend-observer/
├── skill.md                     # Agent execution instructions (English)
├── skill_cn.md                  # Agent execution instructions (Chinese)
├── ONBOARD.md                   # Agent cold-start instructions (English)
├── ONBOARD_CN.md                # Agent cold-start instructions (Chinese)
├── requirements.txt             # Dependency declaration
├── agents/
│   ├── analyzer.md              # PM insight analyzer agent (English)
│   └── analyzer_cn.md           # PM insight analyzer agent (Chinese)
├── scripts/
│   ├── gh_utils.py              # Unified gh CLI utility functions
│   ├── check_rate_limit.py      # API rate limit checker
│   ├── fetch_star_history.py    # Star growth data fetcher
│   ├── radar_pulse.py           # Mode 1 trending fetcher
│   ├── search_repos.py          # Mode 2 search
│   ├── watch_signals.py         # Mode 3 anomaly detection
│   ├── deep_link.py             # Mode 4 relationship analysis
│   ├── generate_report.py       # HTML/MD report generation
│   └── test_oss.py              # Automated tests (6 tiers, 41 tests)
├── config/
│   ├── seed_list.json           # Key developer list
│   └── domain_keywords.json     # Domain keyword mappings
├── templates/
│   ├── radar-pulse.html         # Mode 1 report template (en/cn variants)
│   ├── direction-search.html    # Mode 2 report template
│   ├── signal-watch.html        # Mode 3 report template
│   └── deep-link.html           # Mode 4 report template
├── evals/
│   ├── evals.json               # Test cases (English)
│   └── evals_cn.json            # Test cases (Chinese)
└── references/
    ├── layer_model.md           # Layer classification standard (English)
    └── layer_model_cn.md        # Layer classification standard (Chinese)

Dependencies

Dependency Requirement Check Command
gh CLI >= 2.40.0, authenticated gh auth status
Python >= 3.9 python --version
Extra Python packages None, stdlib only
API quota 5,000 requests/hour when authenticated python scripts/check_rate_limit.py

Common Prerequisites

Must be completed before running any mode:

# 1. Check API quota
python scripts/check_rate_limit.py

Determine the execution strategy based on the returned mode field:

  • full -> Normal execution, including star history fetching
  • degraded -> Skip fetch_star_history.py, use basic data only
  • minimal -> Run search scripts only, skip detail API calls

Mode 1: Proactive Exploration (Radar Pulse)

Trigger: --pulse or "What's worth looking at today?"

Execution Steps

# Step 1: Check quota
python scripts/check_rate_limit.py

# Step 2: Fetch candidates
python scripts/radar_pulse.py --days 7

# Step 3: Read agents/analyzer.md + references/layer_model.md
#         Layer classification -> Filter out L1/L5 -> Select 1-2 with highest PM value

# Step 4: Fetch star history for selected projects (full mode only)
python scripts/fetch_star_history.py owner/repo

Filtering Rules

  1. Label each candidate's Layer
  2. Remove L1 (model-level, too low-level) and L5 (wrapper/demo, noise)
  3. PM value weighting: L2 x 1.5, L3 x 1.3, L4 x 1.0
  4. Take Top 3-5, deep-dive into 1-2

Output Format

# Radar Pulse — {date}
> L2/L3/L4 selection | Filtered {m} from {n} candidates | API: {remaining}/{limit}

## Today's Picks
### {repo} [L?]
> {description}
| Stars | 30d Growth | Language | Created |
|-------|------------|----------|---------|
**Why this one**: {rationale}
**Paradigm signal**: {where the stack is shifting}
**Recommendation**: Deep-dive via Mode 4 / Keep watching

## Also Worth a Look
| Repo | Layer | Stars | One-liner |
|------|-------|-------|-----------|

## Filtered Out
- L1: {n} projects ({examples})
- L5: {n} projects ({examples})

Report saved to: output/radar-pulse_{date}.md


Mode 2: Direction Search

Trigger: User provides a technical direction or keywords

Execution Steps

Step 1: Check Quota

python scripts/check_rate_limit.py

Step 2: Keyword Expansion + Layer 1 Relevance Review

  1. Understand the topic: State the core concept of the user's search in one sentence
  2. Expand keywords: Generate 8-15 search keywords around the topic, covering:
    • Synonymous expressions (swarm -> fleet, colony)
    • Scenario-specific terms (swarm observability, coding agent swarm)
    • Adjacent concepts (coordination, monitoring — things adjacent to swarm)
  3. Layer 1 self-review: Review each keyword; the criterion is "would most results returned be about the same category of thing?" — exact match to every word is not required:
    • Keep: Results are different angles on the same topic
    • Remove: Most results fall into a broader category where the topic is only a small subset
    • Example — topic "agent swarm": swarm orchestration keep (same topic); multi-agent framework remove (swarm is a subset of multi-agent; most results won't be about swarms)
    • Example — topic "Agent-human collaboration": human-in-the-loop agent keep (same topic); AI assistant remove (assistant != human-agent collaboration)
  4. Present to user for confirmation: List kept and removed keywords with rationale; proceed with search only after user confirms

Step 3: Search

python scripts/search_repos.py "{main keyword}" \
  --also "{keyword2}" "{keyword3}" ... \
  --expand "{fallback1}" "{fallback2}" ... \
  --min-stars 20 --min-recall 50

Step 3.5: Dynamic Strategy When Recall Is Low

If deduplicated results < 50, do not silently expand. Instead, present the current situation to the user with three options:

Searched {n} keywords, only {m} unique results after deduplication. Possible reasons and options:

A. This direction hasn't formed a distinct category yet — The relevant capabilities may be embedded as features within larger frameworks rather than existing as standalone projects. Recommend abandoning the search; this finding is itself valuable.

B. Keyword coverage is insufficient — Current keywords may be missing expressions commonly used by the community. I suggest adding the following keywords: {list}. Will continue after confirmation.

C. Proceed with existing results — Although {m} results is a small set, if quality is sufficient, we can go straight to analysis. Suitable for a quick overview of the landscape.

Heuristics for recommending an option:

  • Most keywords return 0 results -> lean toward A (category doesn't exist)
  • Only the main keyword has results, expanded terms return nothing -> lean toward B (missing community terminology)
  • Few results but highly relevant -> lean toward C (small category with clear signal)

Step 4: Layer 2 Result Relevance Classification

After raw search results are returned and before analysis, classify each repo's relevance:

Classification Criteria Handling
high This project is directly working on the topic Include in competitive landscape analysis
medium Related to the topic, but not its primary focus Include based on quality
low Keyword match was coincidental; project is actually about something else Filter out, list under "Filtered Out"

Judgment basis: repo name + description. Ask yourself, "Would this project's author consider themselves working on {user's topic}?"

Step 5: Star History + PM Analysis

# Fetch star growth for key high/medium projects (full mode only)
python scripts/fetch_star_history.py owner/repo

# Read agents/analyzer.md and references/layer_model.md
# Perform Layer classification + PM insight analysis on the data

Output Structure

headline (one paradigm-level judgment)
-> Worth Watching (3-5 deep analysis cards)
-> Competitive Landscape (grouped tables by subcategory; count depends on actual relevant projects)
-> Paradigm Assessment (blue-bordered section)
-> Suggested Deep Dives (3-5, pointing to other Modes)
-> Filtered Out (collapsed, grouped with reasons)

Reports generated in both HTML and MD: output/search_{keyword}_{date}.html/.md


Mode 3: Anomalous Signal Monitoring (Signal Watch)

Trigger: --watch or "Monitor anomalous signals"

Known blind spot: Currently only detects growth anomalies in new projects (created within 90 days). Detecting sudden surges in established projects requires persistent storage for differential comparison — reserved for future iteration.

Execution Steps

Step 1: Check Quota

python scripts/check_rate_limit.py

Step 2: Candidate Discovery

python scripts/watch_signals.py
# Global scan (default), three windows: 7d/30d/90d
# Domain scan: python scripts/watch_signals.py --domain ai-agent
# domain options: ai-agent, llm-tools, ai-infra, mcp, all (default)

The script returns a candidate list (sorted by rough velocity descending), each containing:

  • stars, forks, created, age_days
  • rough_velocity = stars / age_days (rough velocity)
  • fork_ratio = forks / stars (adoption depth signal)

Step 3: Initial Screening + Growth Curve Fetching

  1. Exclude obviously irrelevant items: Check descriptions, exclude games, tutorials, awesome-lists, and other non-technical projects
  2. Fetch star history for remaining candidates (full mode only):
python scripts/fetch_star_history.py owner/repo

Returned growth metrics:

Metric Meaning
avg_daily_7d / avg_daily_30d Average daily growth
acceleration 7d avg / 30d avg; >1 means accelerating
trend_direction Last 3 days avg / prior 4 days avg; indicates current trend
consecutive_growth_days Consecutive days of growth
peak_recency Days since peak; 0 = today
burst_ratio Peak day / 7d avg; high = spike-type growth
recent_7_days[] Daily breakdown; used to assess growth shape

Step 4: Growth Pattern Classification

Examine the shape of recent_7_days[] to determine the growth type:

Pattern Characteristics PM Implication Signal Quality
sustained consecutive > 7 + burst_ratio < 3 Organic growth, real demand High
accelerating trend_direction > 2 + consecutive > 5 Currently surging, act fast Highest
spike+decay burst_ratio > 5 + trend_direction < 0.5 One-time launch burst, likely noise Low
step Single-day spike + stable before and after Event-driven (influencer repost) Medium, watch follow-through

Step 5: Three-Tier Assessment + PM Analysis

Read agents/analyzer.md and assess each candidate holistically:

  • Worth deep-diving: sustained/accelerating pattern + L2/L3 layer
  • Watch: Has growth signal but pattern is unclear, or step-type awaiting follow-through
  • Ignore: spike+decay + L5 wrapper / tutorial / fork_ratio < 0.02

Output Structure

headline (one sentence summarizing the most important signal this period)
-> Signal Overview (table: repo / stars / rough velocity / pattern / assessment)
-> Worth Deep-Diving (3-5 deep cards with growth curve data and PM insights)
-> Watch List (table with brief rationale)
-> Ignored This Period (collapsed, with reasons)

Report saved to: output/signal-watch_{date}.html


Mode 4: Deep Link Analysis

Trigger: User provides a repo URL or owner/repo name

Execution Steps

# Step 1: Check quota
python scripts/check_rate_limit.py

# Step 2: Fetch complete data
python scripts/deep_link.py langchain-ai/langgraph
# URL input supported: python scripts/deep_link.py https://github.com/langchain-ai/langgraph

# Step 3: Fetch star growth curve (full mode only)
python scripts/fetch_star_history.py langchain-ai/langgraph

# Step 4: Read agents/analyzer.md + references/layer_model.md
#         Generate ecosystem map + Layer positioning + paradigm assessment

Output Structure

headline (a bold, tension-bearing judgment that highlights the core tension or most important signal)
-> Basic Profile (table + spark trend chart + commit distribution)
-> Layer Positioning (badge + reasoning + "why not X")
-> Adoption Depth (fork rate / watcher rate / issue activity — distinguishing "spectating" from "actually using")
-> Contributor Structure (table + PM interpretation: bus factor / team vs solo / corporate vs community)
-> Release Cadence (timeline component + product strategy interpretation, not just "how many releases")
-> Issue Composition (table + PM interpretation. If categorization fails (>50% uncategorized),
   must manually sample recent_titles for qualitative analysis as a fallback; never leave blank)
-> Core Innovation (ASCII comparison diagram: traditional approach vs this project's approach.
   This is the fastest path for a PM to understand project value; every report must include one.)
-> Ecosystem Map (ASCII diagram + PM interpretation)
-> Competitor Candidates (collapsed details, annotated with "direct competitor or not" to filter noise)
-> Paradigm Assessment (blue section, structure:
   1. One-sentence paradigm thesis
   2. Core difference: old way vs new way
   3. Who may be threatened
   4. Who is not threatened
   Note: Do not include "relevance to you" — protect privacy)
-> PM Summary (summary-table: maturity / confidence / growth nature / PM value / risk / recommendation)

Output Style

  • CSS uses the --bg/--surface/--border/--accent/--muted variable system, consistent across modes
  • PM insights use .pm-box card component (white background + border), not inline <p>
  • Layer positioning uses .layer-box component with badge + reasoning list + "why not X"
  • Paradigm assessment uses .paradigm component (blue background + border)
  • Competitor candidates go in <details> collapsed sections
  • All technical metrics include plain-language explanations (plain language principle)

Report saved to: output/deep-link_{owner}_{repo}_{date}.html


Seed List Customization

Edit config/seed_list.json to add or remove developers you follow:

{
  "builders": [
    {"github": "username", "note": "why they matter"}
  ],
  "last_updated": "2026-02-18"
}

The default list currently includes 76 important AI builders/orgs, covering 17 categories including labs, agent frameworks, coding agents, inference, platforms, and more.

Reviews (0)

Sign in to write a review.

No reviews yet. Be the first to review!

Comments (0)

Sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Compatible Platforms

Pricing

Free

Related Configs