catalog/repos/199-biotechnologies--claude-deep-research-skill.md

136 lines
5.4 KiB
Markdown
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

# 深度研究报告生成
**GitHub**: [199-biotechnologies/claude-deep-research-skill](https://github.com/199-biotechnologies/claude-deep-research-skill) | **Gitea**: [mc-skills/199-biotechnologies--claude-deep-research-skill](http://192.168.0.109:3000/mc-skills/199-biotechnologies--claude-deep-research-skill)
## 中文摘要
企业级Claude Code研究引擎支持Quick/Standard/Deep/UltraDeep四种模式多来源并行搜索自动验证引用输出含完整参考文献的Markdown/HTML/PDF报告适用于技术选型、复杂课题调研等场景。
## 标签
`研究` `报告生成` `多源搜索` `引用验证` `Claude Code技能`
---
## README 原文
# Deep Research Skill for Claude Code
Enterprise-grade research engine for Claude Code. Produces citation-backed reports with source credibility scoring, multi-provider search, and automated validation.
## Installation
```bash
# Clone into Claude Code skills directory
git clone https://github.com/199-biotechnologies/claude-deep-research-skill.git ~/.claude/skills/deep-research
```
No additional dependencies required for basic usage.
### Optional: search-cli (multi-provider search)
For aggregated search across Brave, Serper, Exa, Jina, and Firecrawl:
```bash
brew tap 199-biotechnologies/tap && brew install search-cli
search config set keys.brave YOUR_KEY # configure at least one provider
```
## Usage
```
deep research on the current state of quantum computing
```
```
deep research in ultradeep mode: compare PostgreSQL vs Supabase for our stack
```
## Research Modes
| Mode | Phases | Duration | Best For |
|------|--------|----------|----------|
| Quick | 3 | 2-5 min | Initial exploration |
| Standard | 6 | 5-10 min | Most research questions |
| Deep | 8 | 10-20 min | Complex topics, critical decisions |
| UltraDeep | 8+ | 20-45 min | Comprehensive reports, maximum rigor |
## Pipeline
Scope → Plan → **Retrieve** (parallel search + agents) → Triangulate → Outline Refinement → Synthesize → Critique (with loop-back) → Refine → Package
Key features:
- **Step 0**: Retrieves current date before searches (prevents stale training-data year assumptions)
- **Parallel retrieval**: 5-10 concurrent searches + 2-3 focused sub-agents returning structured evidence objects
- **First Finish Search**: Adaptive quality thresholds by mode
- **Critique loop-back**: Phase 6 can return to Phase 3 with delta-queries if critical gaps found
- **Multi-persona red teaming**: Skeptical Practitioner, Adversarial Reviewer, Implementation Engineer (Deep/UltraDeep)
- **Disk-persisted citations**: `sources.json` survives context compaction and continuation agents
## Output
Reports saved to `~/Documents/[Topic]_Research_[Date]/`:
- Markdown (primary source of truth)
- HTML (McKinsey-style, auto-opened in browser)
- PDF (professional print via WeasyPrint)
Reports >18K words auto-continue via recursive agent spawning with context preservation.
## Quality Standards
- 10+ sources, 3+ per major claim
- Executive summary 200-400 words
- Findings 600-2,000 words each, prose-first (>=80%)
- Full bibliography with URLs, no placeholders
- Automated validation: `validate_report.py` (9 checks) + `verify_citations.py` (DOI/URL/hallucination detection)
- Validation loop: validate → fix → retry (max 3 cycles)
## Search Tools
| Tool | When | Setup |
|------|------|-------|
| WebSearch | Default, always available | None |
| Exa MCP | Semantic/neural search | MCP config |
| search-cli | Multi-provider aggregation | `brew install search-cli` + API keys |
## Architecture
```
deep-research/
├── SKILL.md # Skill entry point (lean, ~100 lines)
├── reference/
│ ├── methodology.md # 8-phase pipeline details
│ ├── report-assembly.md # Progressive generation strategy
│ ├── quality-gates.md # Validation standards
│ ├── html-generation.md # McKinsey HTML conversion
│ ├── continuation.md # Auto-continuation protocol
│ └── weasyprint_guidelines.md # PDF generation
├── templates/
│ ├── report_template.md # Report structure template
│ └── mckinsey_report_template.html # HTML report template
├── scripts/
│ ├── validate_report.py # 9-check structure validator
│ ├── verify_citations.py # DOI/URL/hallucination checker
│ ├── source_evaluator.py # Source credibility scoring
│ ├── citation_manager.py # Citation tracking
│ ├── md_to_html.py # Markdown to HTML converter
│ ├── verify_html.py # HTML verification
│ └── research_engine.py # Core orchestration engine
└── tests/
└── fixtures/ # Test report fixtures
```
## Version History
| Version | Date | Changes |
|---------|------|---------|
| 2.3.1 | 2026-03-19 | Template/validator harmonization, structured evidence, critique loop-back, multi-persona red teaming |
| 2.3 | 2026-03-19 | Contract harmonization, search-cli integration, dynamic year detection, disk-persisted citations, validation loops |
| 2.2 | 2025-11-05 | Auto-continuation system for unlimited length |
| 2.1 | 2025-11-05 | Progressive file assembly |
| 1.0 | 2025-11-04 | Initial release |
## License
MIT - modify as needed for your workflow.