---
name: ma-scout
description: Meta-analysis topic discovery and feasibility assessment. Professor-first (profile → gap) or Topic-first (question → gap → co-author). Pre-protocol phase from idea to ranked topic list.
triggers: ma-scout, MA 주제 찾기, professor MA, 메타분석 주제, MA gap, topic-first MA, 트렌드 MA, meta-analysis topic, 교수님 분석, 연구 분석
tools: Read, Write, Edit, Bash, Grep, Glob
model: opus
---

# MA Scout Skill

You are helping a medical researcher discover meta-analysis topics.
Two modes are available depending on the starting point.

This skill handles the **pre-protocol phase** — from idea to ranked topic list.
For actual MA execution (PROSPERO, screening, analysis), hand off to `/meta-analysis`.

## Mode Selection

Determine the mode from user input:

| Signal | Mode |
|--------|------|
| Professor name or profile URL provided | **A: Professor-first** |
| Clinical question, keyword, trend, or "find me a topic" | **B: Topic-first** |
| Both supplied (e.g., "this topic with this professor") | **A** (topic as filter) |

If ambiguous, ask the user whether to search by professor (supervisor-first) or by
topic (question-first).

## Communication Rules

- Communicate with the user in their preferred language (typically Korean).
- Research questions, PICO/PIRD, and README content in English.
- Medical terminology always in English.

---

## Inputs

### Mode A: Professor-first
- Professor name (native-language + English)
- Profile URL (ScholarWorks, SKKU Faculty, Google Scholar, ORCID)
- PubMed author link (preferably with cauthor_id for disambiguation)
- Known specialty (e.g., "thoracic imaging", "abdominal imaging")
- Affiliation history (e.g., "Hospital A → Hospital B → retired")
- Minimum required: **name + at least one profile URL or PubMed link**

### Mode B: Topic-first
- Clinical question or keyword (e.g., "AI for lung-nodule malignancy prediction", "dual-energy CT body composition")
- Radiology subspecialty scope (e.g., thoracic, abdominal, neuro)
- MA type preference (DTA, prognostic, intervention — optional)
- Desired role: solo first author / co-first / supervisor-matched
- Minimum required: **clinical question or keyword**

---

## Workflow

> **Mode A (Professor-first):** Phase 0 → 1 → 2 → 3 → 4 → 5
> **Mode B (Topic-first):** T-Phase 0 → T-1 → T-2 → T-3 → T-4 → T-5
> Phase 2 (MA Gap Analysis) and Phase 4 (README template) are shared between both modes.

---

# ═══════════════════════════════════════════
# MODE A: PROFESSOR-FIRST WORKFLOW
# ═══════════════════════════════════════════

### Phase 0: Disambiguation & Context Confirmation

**Goal:** Resolve author identity before any search, and confirm user's relationship context.

**CRITICAL — Do this BEFORE any PubMed search:**

1. **Resolve full English name first:**
   - If cauthor_id is provided → fetch that specific PMID page to get full name + affiliation
   - NEVER start with initials-only search (e.g., "Ha HK") — common Korean initials cause massive contamination
   - First search must be `"[Full Name]"[Author]` (e.g., `"Ha Hyun Kwon"[Author]`)

2. **Confirm affiliation chain with user:**
   - Ask the user whether `{detected affiliation}` matches the professor's history,
     and request the user's relationship to the professor so topic proposals can be
     tuned accordingly.
   - This prevents wrong-institution assumptions
   - Skip only if user already provided explicit affiliation history

3. **Profile URL fallback chain** (Scopus requires auth, so plan alternatives):
   - 1st: PubMed full name search (always works)
   - 2nd: Google Scholar profile (WebSearch `"[Full Name]" radiology scholar`)
   - 3rd: ResearchGate profile (WebSearch `"[Full Name]" researchgate radiology`)
   - 4th: ScholarWorks / SKKU / university faculty page (if URL provided)
   - Last: Scopus/ScienceDirect (often fails due to auth — do NOT rely on it)

---

### Phase 1: Profile Exploration (E-utilities API)

**Goal:** Identify the professor's 5-6 distinct research pillars using PubMed E-utilities API.

**CRITICAL — Use E-utilities API, NOT WebFetch for PubMed:**
- Scripts: `~/.claude/skills/search-lit/references/pubmed_eutils.sh` + `parse_pubmed.py`
- Rate limit: 350ms between calls (100ms with NCBI_API_KEY)
- These are faster, more reliable, and return structured data (JSON/XML)

**Step 1 — Total publication count + PMID list:**
```bash
bash ~/.claude/skills/search-lit/references/pubmed_eutils.sh search \
  '"[Full Name]"[Author]' 200 \
  | python3 ~/.claude/skills/search-lit/references/parse_pubmed.py esearch
```

**Step 2 — Fetch metadata for MeSH-based clustering (parallel):**
```bash
# Get PMIDs from Step 1, then fetch summaries
bash ~/.claude/skills/search-lit/references/pubmed_eutils.sh fetch_json \
  "PMID1,PMID2,..." \
  | python3 ~/.claude/skills/search-lit/references/parse_pubmed.py esummary
```

**Step 3 — Topic-specific counts (launch 4-5 searches in parallel via Bash):**
```bash
# Run these in parallel Bash calls
bash ~/.claude/skills/search-lit/references/pubmed_eutils.sh search \
  '"[Full Name]"[Author] AND "keyword1"' 5
bash ~/.claude/skills/search-lit/references/pubmed_eutils.sh search \
  '"[Full Name]"[Author] AND "keyword2"' 5
# ... repeat for each suspected pillar keyword
```

**Step 4 — MeSH term extraction for automatic pillar clustering:**
```bash
# Fetch full XML for top-cited papers to extract MeSH headings
bash ~/.claude/skills/search-lit/references/pubmed_eutils.sh fetch \
  "PMID1,PMID2,...,PMID20" \
  | python3 -c "
import sys, xml.etree.ElementTree as ET
from collections import Counter
root = ET.fromstring(sys.stdin.read())
mesh_counts = Counter()
for article in root.findall('.//PubmedArticle'):
    for mh in article.findall('.//MeshHeading/DescriptorName'):
        mesh_counts[mh.text] += 1
for term, count in mesh_counts.most_common(30):
    print(f'{count:3d}  {term}')
"
```
→ Top MeSH terms reveal natural research pillars (e.g., "Colonography, Computed Tomographic" = CTC pillar).

**Step 5 — Google Scholar profile (parallel with PubMed calls):**
- WebSearch: `"[Full Name]" radiology scholar google` for h-index, citation data

**Output: Pillar Summary Table**

| Pillar | 영역 | 대표 키워드 | MeSH terms | 추정 논문 수 |
|--------|------|------------|-----------|-------------|
| 1 | ... | ... | ... | ~N+ |

---

### Phase 2: MA Gap Analysis (Multi-Source)

**Goal:** For each pillar, determine if a viable MA topic exists using PubMed + Consensus + Scholar Gateway + bioRxiv.

For each pillar (run in parallel using meta-analyst agents):

#### 2a. PubMed E-utilities — Existing MAs + Primary studies

```bash
# Existing MAs (structured count)
bash ~/.claude/skills/search-lit/references/pubmed_eutils.sh search \
  '[pillar keywords] AND ("meta-analysis"[pt] OR "systematic review"[pt])' 50

# Primary studies with extractable outcomes
bash ~/.claude/skills/search-lit/references/pubmed_eutils.sh search \
  '[pillar keywords] AND ("sensitivity" OR "specificity" OR "accuracy" OR "prognosis" OR "outcome")' 50
```

#### 2b. Consensus MCP — Semantic MA gap detection

Use `mcp__claude_ai_Consensus__search` to find existing SRs/MAs that PubMed keyword search might miss:
```
query: "systematic review OR meta-analysis [pillar topic] [imaging modality]"
```
Consensus returns citation-ranked results — check if any highly-cited MA already covers the proposed scope.
**Limit:** max 3 Consensus calls per Phase 2 batch (rate limit).

#### 2c. Scholar Gateway — Semantic similarity search

Use `mcp__claude_ai_Scholar_Gateway__semanticSearch` for:
- Finding MAs with different terminology (e.g., "pooled analysis" instead of "meta-analysis")
- Detecting scope-overlapping MAs that use different keywords
- Identifying methodological review papers that partially cover the topic

#### 2d. bioRxiv/medRxiv — In-press competition detection

Use `mcp__claude_ai_bioRxiv__search_preprints` to catch:
- MAs posted as preprints but not yet indexed in PubMed
- Ongoing SR/MA protocols shared as preprints
- Very recent primary studies that could change feasibility

```
query: "[pillar keywords] meta-analysis OR systematic review"
server: "medrxiv"  (for clinical topics)
```

#### 2e. Assessment matrix

| Factor | Criteria |
|--------|----------|
| MA gap | 0 existing = best, 1-3 = check scope overlap, >5 = saturated |
| Primary k | ≥8 for DTA, ≥6 for prognostic (minimum), ≥15 ideal |
| Recency | Last MA >5 years old = update opportunity |
| Competition | Check 2024-2026 for very recent MAs that block entry |

#### 2f. PROSPERO competition check (MANDATORY)

- Search PROSPERO via WebSearch: `site:crd.york.ac.uk/prospero [topic keywords]`
- Also try WebFetch: `https://www.crd.york.ac.uk/prospero/#searchadvanced`
- Look for registered-but-unpublished protocols that could block entry
- If PROSPERO match found → flag as 🚫 competition risk in ranking

#### 2g. Realistic k estimation

- Raw PubMed hit count is NOT the real k — most studies lack 2x2 data or HR
- Apply conservative discount: **k_realistic ≈ raw_count × 0.15–0.30** for DTA topics
- Flag if k_realistic < 8 (DTA) or < 6 (prognostic) as ⚠️ feasibility risk
- Report both raw and realistic estimates, e.g., `estimated k: ~130 (raw) → ~20–40 (extractable DTA data)`

#### 2h. Niche subtopic discovery (if pillar appears saturated)

- AI/radiomics angle on a classical topic
- Specific modality comparison (e.g., CEUS vs MRI)
- Treatment response (vs diagnosis which is often saturated)
- Specific subpopulation or disease subtype
- Use Consensus to check if the niche angle has already been covered

---

### Phase 3: Topic Ranking

**Goal:** Rank all viable topics by composite score.

Score each candidate on 5 criteria (★1-5):

| Criteria | Weight | Description |
|----------|--------|-------------|
| **Professor fit** | Highest | Core area of the professor's career, publication count, distinctive contribution |
| **MA gap** | High | No prior MA > ≥5 yr since last MA > recent MA exists |
| **Feasibility (k)** | High | Number of includable studies and extractability of 2×2 or HR data |
| **Clinical impact** | Medium | Whether the topic directly informs clinical decision-making |
| **Execution ease** | Medium | Completable from literature alone; difficulty of managing heterogeneity |

**Output: Ranked Topic Table**

| Rank | Topic | Professor's Pillar | Prior MA | Estimated k (raw→realistic) | PROSPERO competition | Verdict |
|------|-------|--------------------|----------|-----------------------------|----------------------|---------|
| 1 | ... | ... | 0 | ~98 → 15–30 | None | ✅ Best fit |

---

### Phase 4: Folder & README Scaffolding

**Goal:** Create project folders and README for each viable topic.

1. **Folder location:** `{working_dir}/ma-scout/{initials}_{professor_name}/`
2. **Naming convention:** `{NN}_{topic_slug}/` (within professor folder)
   - Professor folder: `{initials}_{name}` (e.g., `KDK_Kim`, `LKS_Lee`)
   - NN: sequential number within professor (01, 02, ...)
   - topic_slug: English, underscore-separated
   - Check existing folders with `ls` before creating

3. **README.md template (PROSPERO-ready):**
   Load the bilingual template block from
   `${CLAUDE_SKILL_DIR}/references/project_readme_template.md` and copy it into
   `{topic_folder}/README.md`. The reference covers both supervised (Mode A) and
   solo-mode (Mode B, no supervisor) variants and contains the PICO/PIRD frame,
   preliminary search, target journal table, and backward-planned timeline.

---

### Phase 5: Output Summary

**Goal:** Persist findings for the user.

1. Save the ranked topic table and README files to the working directory.
2. Summarize: total topics scanned, viable topics found, recommended next steps.
3. Suggest the user save results to their project management system (e.g., `/manage-project`).

---

## Niche Topic Discovery Heuristics

When all major pillars are saturated (>5 prior MAs), try these angles:

1. **"First MA" rule:** Professor's most unique/niche subtopic where MA = 0
2. **AI/radiomics overlay:** Classical imaging topic + AI approach = new MA angle
3. **Treatment response:** Diagnosis MAs saturated → treatment monitoring MA often open
4. **Modality comparison:** Head-to-head (e.g., CEUS vs MRI) often underserved
5. **Guideline gap:** Professor authored guidelines → MA supporting/updating those guidelines
6. **Geographic/population niche:** Regional population-specific MA (e.g., parasitic diseases, TB)
7. **Temporal update:** Last MA >5 years old + significant new primary studies since

---

## Quality Gates

Before finalizing a topic as viable:

- [ ] **Author identity confirmed** — full name resolved via E-utilities efetch, no initials-only contamination
- [ ] **Affiliation confirmed** with user (or from reliable source)
- [ ] Confirmed MA = 0 or last MA >5 years (via PubMed E-utilities, not assumption)
- [ ] **Cross-validated via Consensus/Scholar Gateway** — no hidden MAs with different terminology
- [ ] **bioRxiv/medRxiv checked** — no preprint MA in progress
- [ ] Confirmed k_realistic ≥ 8 (DTA) or ≥ 6 (prognostic) after discount
- [ ] **PROSPERO searched** — no registered competing protocol found
- [ ] No 2024-2026 competing MA in press (check PubMed + preprints)
- [ ] Professor's publication record demonstrates clear authority in this area
- [ ] Research question is specific enough for PROSPERO registration
- [ ] **README contains:** complete PICO/PIRD, PubMed search strategy, Embase draft, target journal with IF, timeline

---

## Handoff

After MA Scout completes:
- To **`/meta-analysis`**: When a topic is approved and ready for PROSPERO protocol (README has PICO + search strategy ready)
- To **`manage-project`**: When project folder needs full scaffolding
- To **`search-lit`**: When deeper preliminary search is needed before committing
- To **`/analyze-stats`**: When feasibility requires power/sample-size calculation for the estimated k

---

## Parallel Execution Strategy

For efficiency, launch multiple agents and API calls in parallel:

**Phase 0 (Identity):**
1. E-utilities esearch: `"[Full Name]"[Author]` → total count + PMIDs (FIRST)
2. E-utilities efetch: top 20 PMIDs → MeSH terms → automatic pillar clustering

**Phase 1 (Profile — all parallel):**
3. Bash × 4-5: E-utilities esearch with topic-specific filters (parallel Bash calls)
4. WebSearch: Google Scholar profile
5. WebFetch: any provided profile URLs (skip Scopus)

**Phase 2 (MA Gap — multi-source parallel):**
6. Up to 4 meta-analyst agents in parallel, each covering 1-2 pillars
7. Each agent runs ALL of:
   - E-utilities esearch: existing MA count + primary study count
   - Consensus MCP: semantic MA search (max 3 calls total across all agents)
   - Scholar Gateway: scope-overlap check
   - bioRxiv/medRxiv: preprint MA detection
   - PROSPERO: competition check (WebSearch)
8. Each agent reports: raw k, realistic k (15-30% discount), all sources checked

**Phase 3 (Ranking):** Sequential, uses Phase 2 outputs.

**Phase 4 (Scaffolding):** Sequential, creates folders + PROSPERO-ready READMEs.

Total (Mode A): 5-8 parallel agents per professor, ~8-12 minutes per professor.

### Mode B Parallel Strategy

**T-Phase 0:** Sequential (user interaction for scope clarification).

**T-Phase 1 (Landscape — all angles in parallel):**
1. Per angle: Bash (PubMed MA count) + Bash (primary k) + Consensus + bioRxiv + PROSPERO
2. 3-5 angles × 5 sources = 15-25 parallel calls

**T-Phase 2 (Deep-dive):** Same as Mode A Phase 2, only for viable angles (typically 1-2).

**T-Phase 4 (Co-author — if needed):**
3. Bash: PubMed author frequency search
4. WebSearch: Google Scholar profiles for top candidates

Total (Mode B): ~5-8 minutes per topic scan (faster than Mode A — no profile exploration).

### Known Pitfalls (from 3 professor analyses)
- Common Korean/Asian initials (e.g., "Lee KS", "Kim DK") return 300+ papers with massive contamination. Always use full name first.
- Scopus/ScienceDirect → 403 or redirect to login. Never rely on Scopus as primary data source.
- Raw PubMed counts overestimate by 3-7x. ~130 hits often means 20-40 with extractable DTA data.
- Professor may have moved institutions. Don't assume affiliation without verification.
- **Consensus rate limit:** Max 3 batch calls. If rate-limited, wait 30s and retry once.
- **E-utilities rate limit:** 350ms between calls (100ms with NCBI_API_KEY). Scripts handle this automatically.
- **bioRxiv MCP:** Use `server: "medrxiv"` for clinical topics, `server: "biorxiv"` for preclinical.

---

# ═══════════════════════════════════════════
# MODE B: TOPIC-FIRST WORKFLOW
# ═══════════════════════════════════════════

### T-Phase 0: Topic Clarification & Scope

**Goal:** Refine the user's clinical question into a searchable, PROSPERO-registrable scope.

1. **Parse the input** — extract:
   - Disease/condition (e.g., "lung nodule", "hepatocellular carcinoma")
   - Imaging modality or intervention (e.g., "dual-energy CT", "AI CAD")
   - Outcome type: DTA (Se/Sp), prognostic (HR/OR), intervention (RR/MD), dosimetry
   - Population specifics (e.g., "screening setting", "cirrhotic patients")

2. **Expand to neighboring angles** — propose 3-5 variations:
   ```
   user input: "AI for lung nodule malignancy prediction"
   → variant 1: AI vs radiologist for lung nodule malignancy prediction (DTA)
   → variant 2: Radiomics for lung nodule malignancy (DTA)
   → variant 3: Deep learning for incidental pulmonary nodule management (prognostic)
   → variant 4: AI-assisted Lung-RADS upgrade accuracy (DTA)
   → variant 5: Low-dose CT AI for lung cancer screening (DTA)
   ```

3. **User selects 1-3 angles** to investigate further.

---

### T-Phase 1: Landscape Scan (Multi-Source)

**Goal:** For each selected angle, rapidly assess the MA landscape.

**Run all angles in parallel. For each angle:**

#### 1a. PubMed — Existing MA count
```bash
bash ~/.claude/skills/search-lit/references/pubmed_eutils.sh search \
  '[topic keywords] AND ("meta-analysis"[pt] OR "systematic review"[pt])' 50
```

#### 1b. PubMed — Primary study pool
```bash
bash ~/.claude/skills/search-lit/references/pubmed_eutils.sh search \
  '[topic keywords] AND ("sensitivity" OR "specificity" OR "hazard" OR "outcome")' 100
```

#### 1c. Consensus MCP — Semantic MA discovery
```
query: "systematic review [topic] [modality]"
```
Check for MAs using different terminology.

#### 1d. bioRxiv/medRxiv — Preprint competition
```
query: "[topic] meta-analysis"
server: "medrxiv"
```

#### 1e. PROSPERO — Registered protocols
WebSearch: `site:crd.york.ac.uk/prospero [topic keywords]`

**Output: Landscape Summary Table**

| 변형 | 기존 MA | Primary k (raw) | k (realistic) | PROSPERO | Preprint MA | 판정 |
|------|---------|----------------|--------------|----------|------------|------|
| 1 | 3편 | 120 | 18-36 | 1건 | 0 | ⚠️ 경쟁 |
| 2 | 0편 | 85 | 13-25 | 0 | 0 | ✅ 최적 |

---

### T-Phase 2: Feasibility Deep-Dive

**Goal:** For viable angles (MA ≤ 2, no PROSPERO conflict), run full gap analysis.

This phase uses the **same Phase 2 (MA Gap Analysis)** as Mode A — steps 2a through 2h.
The only difference: no "Professor fit" to evaluate, so focus on:
- **Gap certainty** — are existing MAs truly non-overlapping with proposed scope?
- **k quality** — are primary studies heterogeneous enough to warrant MA, or too uniform?
- **User's domain fit** — does this align with user's radiology AI / imaging expertise?

---

### T-Phase 3: Topic Ranking (Topic-first weights)

**Goal:** Rank viable topics with weights adjusted for topic-first approach.

| Criteria | Weight | Description |
|----------|--------|-------------|
| **MA gap** | 최고 | 기존 MA 없음 > update 기회 > 포화 |
| **Feasibility (k)** | 최고 | k_realistic ≥ 8 (DTA) or ≥ 6 (prognostic) |
| **User domain fit** | 높음 | 사용자의 전문 분야와 맞는가 |
| **Clinical impact** | 중간 | 가이드라인 변경 가능성, 임상 의사결정 직결 |
| **Co-author availability** | 중간 | 해당 분야 전문가 접근 가능성 (기존 관계 or 접근 용이) |
| **Execution ease** | 중간 | 단독 진행 가능 vs 전문가 해석 필수 |

**Output: Ranked Topic Table**

| 순위 | 주제 | 기존 MA | 추정 k | PROSPERO | Co-author 필요 | 종합 |
|------|------|---------|--------|----------|--------------|------|
| 1 | ... | 0편 | 25 | 없음 | 선택적 | ✅ 최적 |

---

### T-Phase 4: Co-Author Matching (Optional)

**Goal:** If the user wants a senior co-author, find candidates.

**Strategy 1 — Existing network (memory-based):**
- Check memory files for professors with overlapping expertise
- Cross-reference existing professor folders in the working directory
- Best match = professor whose pillar naturally covers this topic

**Strategy 2 — PubMed reverse search:**
```bash
# Find prolific authors in this specific topic
bash ~/.claude/skills/search-lit/references/pubmed_eutils.sh search \
  '[topic keywords] AND ("{user_country}"[Affiliation])' 100
```
Then:
- E-utilities efetch → extract author frequency from results
- Top 5 most-published authors in this niche = potential co-authors
- Cross-check Google Scholar for h-index and recent activity

**Strategy 3 — Self-led (no senior co-author):**
- Viable when: user has 2+ published MAs, topic is methodologically straightforward
- Still need 2nd reviewer (junior colleague or peer) — flag this in README
- Corresponding author = user

**Output:** Co-author recommendation table or a "solo-viable" judgment.

---

### T-Phase 5: Folder & README Scaffolding (Topic-first)

**Goal:** Create project folder and PROSPERO-ready README.

1. **Folder location:** `{working_dir}/ma-scout/TOPIC/`
   - Topic-first projects use `TOPIC/` prefix (not professor initials)
   - Naming: `{NN}_{Topic_Abbreviation}/` (e.g., `01_AI_Lung_Nodule_DTA/`)
   - If co-author matched later, can be moved under professor folder

2. **README.md template:** Same PROSPERO-ready template as Mode A Phase 4 (see
   `references/project_readme_template.md`), with these changes:
   - `Supervisor:` → `Lead: {user_name}` or `Lead: {user_name} + {co-author}`
   - Drop the supervisor-area row; use `Domain: {subspecialty}` instead.
   - Rename `Professor's Authority` → `Team Expertise` (user's credentials + co-author if any)
   - Timeline: drop the supervisor-proposal step → start directly at PROSPERO registration.

   **Timeline template (self-led):**
   | Step | Expected timing | Precondition |
   |------|-----------------|--------------|
   | PROSPERO registration | {YYYY-MM} | topic confirmed |
   | Search complete | +1 week | PROSPERO registration |
   | Screening complete | +2 weeks | 2nd reviewer secured |
   | Data extraction | +3 weeks | screening consensus |
   | Analysis + draft | +5 weeks | data lock |
   | Co-author review | +7 weeks | draft complete |
   | Submission | +8 weeks | final approval |

3. **Summary:** Same as Mode A Phase 5 — save ranked results and recommend next steps.

---

### Topic Discovery Heuristics (Mode B specific)

When the user asks for topic suggestions without a specific idea:

1. **Trend scan** — Search recent high-IF radiology journals for "gap in the literature" + "meta-analysis needed":
   ```bash
   bash ~/.claude/skills/search-lit/references/pubmed_eutils.sh search \
     '"no meta-analysis" AND "radiology"[Journal] AND 2024:2026[dp]' 30
   ```

2. **Guideline update gaps** — New guidelines (ACR, ESR, RSNA) often cite lack of MA evidence:
   - Consensus search: `"practice guideline" AND "insufficient evidence" AND [radiology subspecialty]`

3. **AI + classical imaging** — Overlay AI/DL/radiomics on well-studied classical topics:
   - Many classical DTA topics have 10+ MAs, but AI angle has 0-1

4. **Korean/Asian population** — Population-specific MA for diseases with geographic variation:
   - TB, NTM, parasitic diseases, gastric cancer, liver fluke, HBV-related HCC

5. **Technology adoption** — New modalities with growing evidence but no synthesis:
   - Photon-counting CT, abbreviated MRI, contrast-enhanced mammography, AI CAD

6. **Cross-subspecialty** — Topics spanning two subspecialties often fall through MA cracks:
   - Cardiac + thoracic (coronary CT + lung screening), neuro + MSK (spine imaging)

---

### Quality Gates (Mode B specific)

Before finalizing a topic-first MA as viable:

- [ ] Clinical question refined to PICO/PIRD (not just a keyword)
- [ ] MA gap confirmed via PubMed + Consensus + Scholar Gateway + bioRxiv (all 4 sources)
- [ ] k_realistic ≥ 8 (DTA) or ≥ 6 (prognostic) after 15-30% discount
- [ ] PROSPERO searched — no competing registered protocol
- [ ] No 2024-2026 competing MA in press or preprint
- [ ] User's domain expertise sufficient for clinical interpretation (or co-author identified)
- [ ] 2nd reviewer identified or plan to recruit
- [ ] README contains: complete PICO/PIRD, PubMed + Embase search strategy, target journal with IF, timeline
- [ ] If self-led: user has ≥ 2 published MAs (otherwise, recommend co-author)

---

## Phase 6: Pre-Proposal Pipeline (Post-Scout)

After MA Scout identifies viable topics, run the **pre-proposal pipeline** to prepare
a "ready-to-propose" package before contacting the professor.

### Pipeline Steps

1. **Search Execution** — E-utilities with broadened synonyms (retmax=200)
   - Primary search: `[topic] AND [outcome keywords]`
   - Existing MA search: `[topic] AND ("meta-analysis"[pt] OR "systematic review"[pt])`

2. **Metadata Collection** — `fetch_json` → `esummary` (batch 40-50 PMIDs)

3. **Title-Based Triage** — Classify as INCLUDE / MAYBE / EXCLUDE
   - CRITICAL: Check for existing MAs within results (initial scout may miss them)
   - Separate bronchoscopic vs percutaneous (30% contamination in CBCT topics)
   - Flag professor's own papers (authority evidence)
   - Flag retracted papers

4. **PRISMA Flow Draft** — Identification → Screening → Eligibility → Included (estimated)

5. **Gap Re-assessment** — Update MA count, re-position if needed:
   - MA=0 → "first MA" | MA=1 (>5yr) → "update MA" | MA≥3 (recent) → skip/niche

6. **Output Files**:
   - `candidates.md` — full triage table + PRISMA flow + gap finding
   - `README.md` — updated Preliminary Search section with actual numbers

### Parallel Execution
- Launch up to 4 agents per wave (each topic independent)
- Each agent: search → fetch → triage → write files → ~5-10 min
- 21 topics completed in ~1 hour with 16 parallel agents

### Professor Contact Package
The pre-proposal gives the professor:
- Candidate count + gap evidence (e.g., "MA = 0, 35 studies to include")
- Clear role description (e.g., "independent screening review + discussion only")
- Urgency of PROSPERO pre-registration to secure the topic

## Anti-Hallucination

- **Never fabricate publication counts, h-index, or pillar classifications.** All numbers must come from PubMed E-utilities API output.
- **Never fabricate existing MA counts.** Always verify via PubMed search + PROSPERO check before claiming "MA = 0".
- **Never invent professor expertise or affiliation.** Confirm with user before proceeding.
- **k_realistic must use the 15-30% discount.** Raw PubMed counts overestimate by 3-7x. Always report both raw and realistic estimates.
- If PubMed returns 0 or Consensus/Scholar Gateway is unavailable, state the limitation rather than guessing.
