# 深度研究报告生成 **GitHub**: [199-biotechnologies/claude-deep-research-skill](https://github.com/199-biotechnologies/claude-deep-research-skill) | **Gitea**: [mc-skills/199-biotechnologies--claude-deep-research-skill](http://192.168.0.109:3000/mc-skills/199-biotechnologies--claude-deep-research-skill) ## 中文摘要 企业级Claude Code研究引擎,支持Quick/Standard/Deep/UltraDeep四种模式,多来源并行搜索,自动验证引用,输出含完整参考文献的Markdown/HTML/PDF报告,适用于技术选型、复杂课题调研等场景。 ## 标签 `研究` `报告生成` `多源搜索` `引用验证` `Claude Code技能` --- ## README 原文 # Deep Research Skill for Claude Code Enterprise-grade research engine for Claude Code. Produces citation-backed reports with source credibility scoring, multi-provider search, and automated validation. ## Installation ```bash # Clone into Claude Code skills directory git clone https://github.com/199-biotechnologies/claude-deep-research-skill.git ~/.claude/skills/deep-research ``` No additional dependencies required for basic usage. ### Optional: search-cli (multi-provider search) For aggregated search across Brave, Serper, Exa, Jina, and Firecrawl: ```bash brew tap 199-biotechnologies/tap && brew install search-cli search config set keys.brave YOUR_KEY # configure at least one provider ``` ## Usage ``` deep research on the current state of quantum computing ``` ``` deep research in ultradeep mode: compare PostgreSQL vs Supabase for our stack ``` ## Research Modes | Mode | Phases | Duration | Best For | |------|--------|----------|----------| | Quick | 3 | 2-5 min | Initial exploration | | Standard | 6 | 5-10 min | Most research questions | | Deep | 8 | 10-20 min | Complex topics, critical decisions | | UltraDeep | 8+ | 20-45 min | Comprehensive reports, maximum rigor | ## Pipeline Scope → Plan → **Retrieve** (parallel search + agents) → Triangulate → Outline Refinement → Synthesize → Critique (with loop-back) → Refine → Package Key features: - **Step 0**: Retrieves current date before searches (prevents stale training-data year assumptions) - **Parallel retrieval**: 5-10 concurrent searches + 2-3 focused sub-agents returning structured evidence objects - **First Finish Search**: Adaptive quality thresholds by mode - **Critique loop-back**: Phase 6 can return to Phase 3 with delta-queries if critical gaps found - **Multi-persona red teaming**: Skeptical Practitioner, Adversarial Reviewer, Implementation Engineer (Deep/UltraDeep) - **Disk-persisted citations**: `sources.json` survives context compaction and continuation agents ## Output Reports saved to `~/Documents/[Topic]_Research_[Date]/`: - Markdown (primary source of truth) - HTML (McKinsey-style, auto-opened in browser) - PDF (professional print via WeasyPrint) Reports >18K words auto-continue via recursive agent spawning with context preservation. ## Quality Standards - 10+ sources, 3+ per major claim - Executive summary 200-400 words - Findings 600-2,000 words each, prose-first (>=80%) - Full bibliography with URLs, no placeholders - Automated validation: `validate_report.py` (9 checks) + `verify_citations.py` (DOI/URL/hallucination detection) - Validation loop: validate → fix → retry (max 3 cycles) ## Search Tools | Tool | When | Setup | |------|------|-------| | WebSearch | Default, always available | None | | Exa MCP | Semantic/neural search | MCP config | | search-cli | Multi-provider aggregation | `brew install search-cli` + API keys | ## Architecture ``` deep-research/ ├── SKILL.md # Skill entry point (lean, ~100 lines) ├── reference/ │ ├── methodology.md # 8-phase pipeline details │ ├── report-assembly.md # Progressive generation strategy │ ├── quality-gates.md # Validation standards │ ├── html-generation.md # McKinsey HTML conversion │ ├── continuation.md # Auto-continuation protocol │ └── weasyprint_guidelines.md # PDF generation ├── templates/ │ ├── report_template.md # Report structure template │ └── mckinsey_report_template.html # HTML report template ├── scripts/ │ ├── validate_report.py # 9-check structure validator │ ├── verify_citations.py # DOI/URL/hallucination checker │ ├── source_evaluator.py # Source credibility scoring │ ├── citation_manager.py # Citation tracking │ ├── md_to_html.py # Markdown to HTML converter │ ├── verify_html.py # HTML verification │ └── research_engine.py # Core orchestration engine └── tests/ └── fixtures/ # Test report fixtures ``` ## Version History | Version | Date | Changes | |---------|------|---------| | 2.3.1 | 2026-03-19 | Template/validator harmonization, structured evidence, critique loop-back, multi-persona red teaming | | 2.3 | 2026-03-19 | Contract harmonization, search-cli integration, dynamic year detection, disk-persisted citations, validation loops | | 2.2 | 2025-11-05 | Auto-continuation system for unlimited length | | 2.1 | 2025-11-05 | Progressive file assembly | | 1.0 | 2025-11-04 | Initial release | ## License MIT - modify as needed for your workflow.