User Guide
What You Can Do with the Clo-Author
How It Works
You describe what you want in plain English. Claude handles the rest.
| You Do | Happens Automatically |
|---|---|
| Describe what you want | Claude selects and runs the right skills |
| Approve plans | Orchestrator coordinates agents via dependency graph |
| Review final output | Worker-critic pairs ensure quality |
| Say “commit” when ready | Quality gates enforce score thresholds |
For complex or ambiguous tasks, Claude may ask 3–5 clarifying questions to create a requirements specification before planning. Each requirement is tagged MUST/SHOULD/MAY, and each aspect is marked CLEAR/ASSUMED/BLOCKED.
Research Workflow
A typical research project flows through these phases, driven by dependencies (not forced sequence):
Phase 1: Discovery
/discover → Interactive interview → Research Spec
/discover --lit → Literature Synthesis (Librarian + librarian-critic)
/discover --data → Data Assessment (Explorer + explorer-critic)
Phase 2: Strategy
/strategize → Strategy Memo (Strategist + strategist-critic)
Phase 3: Execution
/analyze → Scripts + Output (Data-engineer + Coder + coder-critic)
/analyze --dual r,py → Same in two languages + convergence check
/write → Paper Sections (Writer + humanizer pass)
Phase 4: Peer Review
/review Paper/main.tex → Weighted Score (multi-agent parallel)
/review --peer → domain-referee + methods-referee
Phase 5: Submission
/submit --target → Journal Recommendations
/submit --final → Final Gate (score >= 95)
You can enter at any stage. Use /new-project for the full pipeline. Each skill produces structured output and saves reports to quality_reports/.
Discovery Phase
/discover
Interactive research interview that builds your research specification and domain profile. Equivalent to sitting down with a mentor and sharpening your question.
/discover --lit [topic]
Dispatches the Librarian (collector) and librarian-critic (coverage reviewer). Searches top-5 generals, NBER, field journals, SSRN/RePEc. Produces annotated bibliography, BibTeX entries, frontier map, and positioning recommendation.
/discover --data [question]
Dispatches the Explorer (data finder) and explorer-critic (data quality reviewer). Searches public, administrative, survey, and novel data sources. Scores feasibility (A/B/C/D) and measurement quality.
/discover --ideate [topic]
Generates research questions and identification strategies for a broad topic.
Strategy Phase
/strategize
Dispatches the Strategist (proposer) and strategist-critic (reviewer). Produces strategy memo, pseudo-code, robustness plan, and falsification tests. The strategist-critic validates your assumptions, checks inference, and stress-tests the design.
/strategize --pap
Dispatches the Strategist in PAP mode. Drafts in AEA/OSF/EGAP format with outcomes, subgroups, multiple testing corrections, and power calculations.
Execution Phase
/analyze [dataset]
Dispatches the Data-engineer (cleaning + visualization), Coder (analysis implementation), and coder-critic (reviewer). End-to-end analysis: data cleaning, main specification, robustness checks, publication-ready tables and figures. Supports R, Stata, Python, and Julia.
Use --dual r,python to run the same analysis in two languages simultaneously. Both implementations follow the same strategy memo; results are compared within tolerance. Divergences are flagged. See the Command Reference for details.
/write [section]
Dispatches the Writer agent. Drafts paper sections with proper economics structure:
- Introduction: contribution paragraph within first 2 pages, effect sizes stated, clear identification preview
- Empirical Strategy: per-design template (DiD, IV, RDD, etc.) with assumption discussion
- Results: proper table/figure references, statistical vs economic significance
- Notation protocol: \(Y_{it}\), \(D_{it}\), \(ATT(g,t)\) — consistent throughout
Anti-hedging rules enforced. Humanizer pass automatically strips 24 AI writing patterns.
Review Phase
/review [file]
Smart routing based on file type and flags:
| Input | What Runs |
|---|---|
/review Paper/main.tex |
writer-critic + strategist-critic + Verifier (comprehensive) |
/review scripts/R/analysis.R |
coder-critic standalone |
/review --peer |
domain-referee + methods-referee (simulated peer review) |
/review --peer JHR |
Same, calibrated to JHR’s review culture |
/review --methods Paper/main.tex |
strategist-critic standalone |
/review --proofread Paper/main.tex |
writer-critic standalone |
/review --code scripts/R/ |
coder-critic standalone (code quality only) |
/review --replicate python |
Re-implement existing R code in Python, compare outputs |
/review --all |
All critics in parallel + weighted aggregate score |
/review --peer [journal]
Dispatches two specialized blind referees that mimic real journal review:
- domain-referee: Expert in your paper’s subject literature. Checks contributions, positioning, missing citations, substantive arguments, external validity.
- methods-referee: Expert in econometric methods. Checks identification strategy, estimation, inference, robustness, replication readiness.
Both referees are independent (neither sees the other’s report). The Orchestrator synthesizes the editorial decision (Accept/Minor/Major/Reject).
Journal calibration: Add a journal name to calibrate both referees to that journal’s review culture. For example, /review --peer JHR makes the referees prioritize policy relevance and clean identification (what JHR referees actually care about). 15 journals are pre-populated in .claude/rules/journal-profiles.md; you can add your own. Without a journal name, referees review as generic top-field reviewers.
See the full command reference for all flags and subcommands.
Revision Phase
/revise [report]
Classifies each referee comment and routes to the right agent:
| Class | Routing | Action |
|---|---|---|
| NEW ANALYSIS | → Coder | Flag for user, create analysis task |
| CLARIFICATION | → Writer | Draft revised text |
| DISAGREE | → User | Draft diplomatic pushback (flagged for review) |
| MINOR | → Writer | Draft fix directly |
Produces point-by-point response letter with diplomatic disagreement protocol.
Presentations
/talk [format]
Dispatches the Storyteller (creator) and storyteller-critic (talk reviewer). Generates Beamer .tex talks derived from your paper in 4 formats:
| Format | Slides | Duration | Content |
|---|---|---|---|
| Job Market | 40–50 | 45–60 min | Full story, all results, mechanism, robustness |
| Seminar | 25–35 | 30–45 min | Motivation, main result, 2 robustness checks |
| Short | 10–15 | 15 min | Question, method, key result, implication |
| Lightning | 3–5 | 5 min | Hook, result, so-what |
Talk scores are advisory (non-blocking).
Subcommands: /talk create [format], /talk audit [file], /talk compile [file].
Submission Phase
/submit --target
Journal targeting. Produces ranked journal list using domain-profile tiers, formatting requirements, and submission strategy.
/submit --package
Build replication package: AEA-format README, master script, numbered script order, deposit checklist.
/submit --audit
Dispatches the Verifier in submission mode (all 10 checks):
- LaTeX compilation
- Script execution
- File integrity
- Output freshness
- Package inventory
- Dependency verification
- Data provenance
- End-to-end execution
- Output cross-reference
- README completeness (AEA format)
/submit --final
Final submission gate. Requires aggregate score >= 95 with all components >= 80. Generates cover letter draft and submission checklist.
Utilities
/tools [subcommand]
| Subcommand | What It Does |
|---|---|
/tools commit [msg] |
Stage, commit, PR, merge |
/tools compile [file] |
3-pass XeLaTeX + BibTeX |
/tools validate-bib |
Cross-reference citations |
/tools journal |
Research journal timeline |
/tools learn |
Extract session discoveries into persistent skills |
/tools deploy |
Build and deploy to GitHub Pages |
Quality Gates
Every file gets a quality score from 0 to 100, computed as a weighted aggregate across components. The orchestrator manages this automatically.
| Score | Gate | Applies To |
|---|---|---|
| 95+ | Submission | Aggregate + all components >= 80 |
| 90+ | PR | Weighted aggregate (blocking) |
| 80+ | Commit | Weighted aggregate (blocking) |
| < 80 | Blocked | Must fix before committing |
| – | Advisory | Talks (reported, non-blocking) |
See the full scoring formula in the Architecture Reference.
Exploration Workflow
The explorations/ folder provides a structured sandbox for experimental work.
The problem: Without structure, experimental code scatters across the repository.
The solution: All experimental work goes into explorations/ first:
explorations/
├── [active-project]/
│ ├── README.md # Goal, hypotheses, status
│ ├── R/ # Code iterations (_v1, _v2)
│ └── output/ # Results
└── ARCHIVE/
├── completed_[name]/ # Graduated to production
└── abandoned_[name]/ # Documented why stopped
Fast-Track vs Plan-First
| Question | Answer | Workflow |
|---|---|---|
| “Will this ship?” | YES | Plan-First (80/100 quality) |
| “Am I testing an idea?” | YES | Fast-Track (60/100 quality) |
| “Does this improve the project?” | NO | Don’t build it |
Self-Improvement
Skill Extraction: /tools learn
For discoveries that deserve more than a one-line tag, use /tools learn to create a full skill.
| Situation | Use |
|---|---|
| One-liner fix | [LEARN:category] tag in MEMORY.md |
| Multi-step workflow | /tools learn to create full skill |
| Error + root cause + solution | /tools learn if reusable, [LEARN] if not |
Customizing for Your Domain
Domain Profile
The domain profile (.claude/rules/domain-profile.md) calibrates all agents to your field. Fill it in manually or use /discover (interactive interview) to populate it:
- Field & adjacent subfields — inferred from your topic
- Target journals — ranked by tier
- Common data sources — datasets typical for your area
- Common identification strategies — designs used in your literature
- Field conventions — estimation quirks, clustering norms
- Seminal references — papers every referee expects you to cite
- Referee concerns — the “gotcha” questions referees always ask
Creating Custom Skills
Create a skill when you repeatedly explain the same 3+ step workflow to Claude. Each skill lives in .claude/skills/[name]/SKILL.md:
---
name: your-skill-name
description: [What it does] + [When to use] + [Key capabilities]
argument-hint: "[brief hint for user]"
allowed-tools: ["Read", "Write", "Edit", "Bash", "Task"]
---Quick start: Copy templates/skill-template.md to .claude/skills/your-skill-name/SKILL.md and customize.
See the full command reference in the Customization Guide.
Troubleshooting
LaTeX Won’t Compile
Symptom: xelatex errors or missing packages.
- Check you have XeLaTeX installed:
which xelatex - Ensure
TEXINPUTSincludesPreambles/: the/tools compileskill handles this - Missing package? Install via TeX Live:
tlmgr install [package]
Hooks Not Firing
Symptom: No file protection or context survival.
- Check hooks are configured in
.claude/settings.json - Ensure Python 3 is available:
which python3 - Check hook file permissions:
ls -la .claude/hooks/
Claude Ignores Rules
Symptom: Claude doesn’t follow conventions in .claude/rules/.
- Rules use
paths:frontmatter — check the path matches your files - Too many rules? Claude follows ~150 instructions reliably. Consolidate.
- Try: “Read
.claude/rules/[rule].mdand follow it for this task”
Context Lost After Compaction
Symptom: Claude forgets what you were working on.
- Point Claude to the plan: “Read
quality_reports/plans/[latest].md” - Check session log: “Read
quality_reports/session_logs/[latest].md” - The
post-compact-restore.pyhook should print recovery info automatically
Quality Score Too Low
Symptom: Score stuck below 80, can’t commit.
- Run
/review --allto get detailed issue breakdown - Fix critical issues first (they cost the most points)
- Check per-component scores — one weak component can drag down the aggregate