Web Searcher
Autonomous web research agent that performs multi-step searches, follows links, extracts data, and synthesizes findings into structured reports. Use when asked to research a topic, find information ac
Description
name: web-searcher description: Autonomous web research agent that performs multi-step searches, follows links, extracts data, and synthesizes findings into structured reports. Use when asked to research a topic, find information across multiple sources, compare options, gather market data, compile lists, or answer questions requiring deep web investigation beyond a single search.
Web Searcher Agent
Workflow
-
Parse the query — Break the user's request into 2-5 specific search queries that cover different angles of the topic.
-
Search phase — Execute searches using
web_search. Rate limit: max 3 searches, then assess before continuing. -
Deep dive phase — For promising results, use
web_fetchto extract full content. Prioritize:- Primary sources over aggregators
- Recent content over old (check dates)
- Authoritative domains over random blogs
-
Cross-reference — Compare findings across sources. Flag contradictions. Note consensus.
-
Synthesize — Compile findings into a clear, structured response with:
- Key findings (bullet points)
- Sources cited (URLs)
- Confidence level (high/medium/low per claim)
- Gaps identified (what couldn't be found)
Search Strategies
Factual queries
Search → verify across 2+ sources → report with citations.
Comparison/market research
Search each option separately → fetch detail pages → build comparison table → recommend.
People/company research
Search name + context → fetch LinkedIn/company pages → cross-reference news → compile profile.
How-to/technical
Search with specific technical terms → fetch documentation/guides → distill steps.
Guidelines
- Max 10 searches per task to avoid rate limits and token waste.
- Max 5 page fetches — be selective about which URLs to deep-dive.
- Always include source URLs so the user can verify.
- If a search returns nothing useful, rephrase and retry once before moving on.
- For time-sensitive info, use
freshnessparameter (pd/pw/pm/py). - Prefer
web_fetchwithmaxChars: 5000to keep context manageable. - If the task is massive, suggest breaking it into sub-tasks or spawning sub-agents.
Output Format
## [Topic]
### Key Findings
- Finding 1 (Source: url)
- Finding 2 (Source: url)
### Details
[Expanded analysis]
### Sources
1. [Title](url) — what was found here
2. [Title](url) — what was found here
### Confidence & Gaps
- High confidence: [claims well-supported]
- Low confidence: [claims with limited sources]
- Not found: [what couldn't be determined]
Reviews (0)
No reviews yet. Be the first to review!
Comments (0)
No comments yet. Be the first to share your thoughts!