🎁 Get the FREE AI Skills Starter Guide β€” Subscribe β†’
BytesAgainBytesAgain
πŸ¦€ ClawHub

Deep Research

by @b143kc47

use for adaptive deep research, broad but accurate information gathering, literature review, github and project due diligence, source graph investigation, ci...

Versionv1.0.1
Downloads67
πŸ’‘ Examples

1. Identify the deliverable: direct answer, research memo, literature review, project comparison, due diligence, timeline, implementation recommendation, or full cited report. 2. Choose effort based on risk and ambiguity: - quick: 2-4 meaningful hops, 2+ source classes, for low-risk checks. - standard: 5-8 hops, 3+ source classes, for normal research. - deep: 9-14 hops, 4+ source classes, for broad synthesis. - exhaustive: 15+ hops or user-specified budget, 5+ source classes, for hard, contested, or high-stakes research. 3. Initialize a run:

python {baseDir}/scripts/research_ledger.py init \
  --question "" \
  --out-dir research_runs \
  --effort deep \
  --deliverable "evidence-backed research memo"

4. Load research-protocol.md for the workflow and query-playbook.md for search patterns. 5. After each meaningful retrieval, source opening, repo inspection, citation traversal, or verification step, log a hop. After each source contributes a reusable claim, log evidence. 6. Before finalizing, run:

python {baseDir}/scripts/research_ledger.py lint --run-dir 

7. Use report-template.md. Cite evidence IDs such as [E0001] for high-impact claims.

View on ClawHub
TERMINAL
clawhub install b143kc47-deep-research

πŸ§ͺ Use this skill with your agent

Most visitors already have an agent. Pick your environment, install or copy the workflow, then run the smoke-test prompt above.

πŸ” Can't find the right skill?

Search 60,000+ AI agent skills β€” free, no login needed.

Search Skills β†’