catalog/repos/199-biotechnologies--claude-deep-research-skill.md

136 lines
5.4 KiB
Markdown
Raw Normal View History

2026-04-06 15:47:33 +08:00
# 深度研究报告生成
**GitHub**: [199-biotechnologies/claude-deep-research-skill](https://github.com/199-biotechnologies/claude-deep-research-skill) | **Gitea**: [mc-skills/199-biotechnologies--claude-deep-research-skill](http://192.168.0.109:3000/mc-skills/199-biotechnologies--claude-deep-research-skill)
## 中文摘要
企业级Claude Code研究引擎支持Quick/Standard/Deep/UltraDeep四种模式多来源并行搜索自动验证引用输出含完整参考文献的Markdown/HTML/PDF报告适用于技术选型、复杂课题调研等场景。
## 标签
`研究` `报告生成` `多源搜索` `引用验证` `Claude Code技能`
---
## README 原文
# Deep Research Skill for Claude Code
Enterprise-grade research engine for Claude Code. Produces citation-backed reports with source credibility scoring, multi-provider search, and automated validation.
## Installation
```bash
# Clone into Claude Code skills directory
git clone https://github.com/199-biotechnologies/claude-deep-research-skill.git ~/.claude/skills/deep-research
```
No additional dependencies required for basic usage.
### Optional: search-cli (multi-provider search)
For aggregated search across Brave, Serper, Exa, Jina, and Firecrawl:
```bash
brew tap 199-biotechnologies/tap && brew install search-cli
search config set keys.brave YOUR_KEY # configure at least one provider
```
## Usage
```
deep research on the current state of quantum computing
```
```
deep research in ultradeep mode: compare PostgreSQL vs Supabase for our stack
```
## Research Modes
| Mode | Phases | Duration | Best For |
|------|--------|----------|----------|
| Quick | 3 | 2-5 min | Initial exploration |
| Standard | 6 | 5-10 min | Most research questions |
| Deep | 8 | 10-20 min | Complex topics, critical decisions |
| UltraDeep | 8+ | 20-45 min | Comprehensive reports, maximum rigor |
## Pipeline
Scope → Plan → **Retrieve** (parallel search + agents) → Triangulate → Outline Refinement → Synthesize → Critique (with loop-back) → Refine → Package
Key features:
- **Step 0**: Retrieves current date before searches (prevents stale training-data year assumptions)
- **Parallel retrieval**: 5-10 concurrent searches + 2-3 focused sub-agents returning structured evidence objects
- **First Finish Search**: Adaptive quality thresholds by mode
- **Critique loop-back**: Phase 6 can return to Phase 3 with delta-queries if critical gaps found
- **Multi-persona red teaming**: Skeptical Practitioner, Adversarial Reviewer, Implementation Engineer (Deep/UltraDeep)
- **Disk-persisted citations**: `sources.json` survives context compaction and continuation agents
## Output
Reports saved to `~/Documents/[Topic]_Research_[Date]/`:
- Markdown (primary source of truth)
- HTML (McKinsey-style, auto-opened in browser)
- PDF (professional print via WeasyPrint)
Reports >18K words auto-continue via recursive agent spawning with context preservation.
## Quality Standards
- 10+ sources, 3+ per major claim
- Executive summary 200-400 words
- Findings 600-2,000 words each, prose-first (>=80%)
- Full bibliography with URLs, no placeholders
- Automated validation: `validate_report.py` (9 checks) + `verify_citations.py` (DOI/URL/hallucination detection)
- Validation loop: validate → fix → retry (max 3 cycles)
## Search Tools
| Tool | When | Setup |
|------|------|-------|
| WebSearch | Default, always available | None |
| Exa MCP | Semantic/neural search | MCP config |
| search-cli | Multi-provider aggregation | `brew install search-cli` + API keys |
## Architecture
```
deep-research/
├── SKILL.md # Skill entry point (lean, ~100 lines)
├── reference/
│ ├── methodology.md # 8-phase pipeline details
│ ├── report-assembly.md # Progressive generation strategy
│ ├── quality-gates.md # Validation standards
│ ├── html-generation.md # McKinsey HTML conversion
│ ├── continuation.md # Auto-continuation protocol
│ └── weasyprint_guidelines.md # PDF generation
├── templates/
│ ├── report_template.md # Report structure template
│ └── mckinsey_report_template.html # HTML report template
├── scripts/
│ ├── validate_report.py # 9-check structure validator
│ ├── verify_citations.py # DOI/URL/hallucination checker
│ ├── source_evaluator.py # Source credibility scoring
│ ├── citation_manager.py # Citation tracking
│ ├── md_to_html.py # Markdown to HTML converter
│ ├── verify_html.py # HTML verification
│ └── research_engine.py # Core orchestration engine
└── tests/
└── fixtures/ # Test report fixtures
```
## Version History
| Version | Date | Changes |
|---------|------|---------|
| 2.3.1 | 2026-03-19 | Template/validator harmonization, structured evidence, critique loop-back, multi-persona red teaming |
| 2.3 | 2026-03-19 | Contract harmonization, search-cli integration, dynamic year detection, disk-persisted citations, validation loops |
| 2.2 | 2025-11-05 | Auto-continuation system for unlimited length |
| 2.1 | 2025-11-05 | Progressive file assembly |
| 1.0 | 2025-11-04 | Initial release |
## License
MIT - modify as needed for your workflow.