Stop using AI as a chatbot. Start building it into your research infrastructure. A presentation from a researcher showed how Claude Code transforms from a conversation partner into a persistent, multi-agent research system: CLAUDE.md as project memory, Skills as domain expert specializations, and parallel AI agent teams that turn a single researcher into the coordinator of an AI intelligence squad.
| *Source: ClaudeCodeTools Presentation (PDF) | η±ε―ε―-η±ηζ΄» Weibo analysis* |
The Three Layers
Layer 3: AI Intelligence Team
Multiple agents working in parallel
Researcher = coordinator, not executor
β
Layer 2: Domain Expert Personas
Skills, Commands, Personas
"Reviewer #2", "Editor", "Stats Checker"
β
Layer 1: Persistent Project Memory
CLAUDE.md = project manual
Goals, decisions, data locations, code standards
Layer 1: CLAUDE.md as Research Memory
Every time Claude Code starts, it reads CLAUDE.md β your projectβs persistent, editable memory:
| What CLAUDE.md Stores | Why It Matters |
|---|---|
| Project goals & hypotheses | AI understands why, not just what |
| Key decisions made | No re-explaining past choices |
| Data file locations | Direct access to your datasets |
| Code standards & conventions | Consistent output across sessions |
| Current progress & open questions | Picks up where you left off |
Without CLAUDE.md, AI is a forgetful goldfish β brilliant but amnesiac. With it, AI becomes a colleague who remembers everything about your project.
Layer 2: Domain Expert Personas
Through Skills, Commands, and Personas, you can weaponize Claude as different specialists:
| Persona | What It Does |
|---|---|
| βReviewer #2β | The harshest, most meticulous critic β scrutinizes your code, logic, and methodology with the rigor of a hostile peer reviewer |
| βEditorβ | Polishes awkward prose into flowing, readable academic writing |
| βStats Checkerβ | Validates statistical methods, checks assumptions, flags p-hacking risks |
| βLiterature Scoutβ | Finds related work, identifies gaps, suggests citations |
| βViz Designerβ | Creates publication-quality figures from your data |
Youβre not asking Claude to βbeβ these people β youβre loading specific instruction sets that activate different reasoning modes and quality standards.
Layer 3: AI Intelligence Team
The most powerful concept: launch multiple independent AI agents simultaneously, each handling a different research task:
βββββββββββββββ
β Researcher β
β(Coordinator)β
ββββββββ¬βββββββ
ββββββββ¬βββββββΌβββββββ¬βββββββ
βΌ βΌ βΌ βΌ βΌ
βββββββββββββββββββββββββββββββββββ
βCode ββData ββStatsββLit ββResultβ
βAuditββ Viz ββCheckββReviewββReportβ
βββββββββββββββββββββββββββββββββββ
β β β β β
ββββββββ΄βββββββΌβββββββ΄βββββββ
βΌ
Consolidated Output
Each agent works in its own context (isolated worktree), produces results, and the researcher consolidates. Humans shift from executors to managers of AI teams.
The Research Workflow Revolution
| Traditional Research | AI-Augmented Research |
|---|---|
| Manually clean data | Agent reads data directly, runs cleanup scripts |
| Print-statement debugging | Agent runs code by your rules, finds bugs |
| Half a day on one slide | Viz agent generates publication figures |
| Week-long literature review | Literature agent scans and synthesizes in hours |
| Solo peer-review prep | βReviewer #2β attacks your paper before submission |
The Uncomfortable Question
The article ends with a provocation:
βIf experimental validation costs, data visualization costs, and literature review costs all approach zero β how much meaning is left in our 8-hour daily βbrick-layingβ?β
This isnβt about replacement. Itβs about reallocation: when AI handles the mechanical research tasks, what do researchers spend their freed time on? The answer should be: thinking harder, asking better questions, and doing the creative work that AI canβt.
Case Study: Building This for Real
I actually built this 3-layer infrastructure for my own research β a formal verification project (relational cell morphing with CHC constraint solving). Hereβs exactly what I set up and how it works in practice.
What I Built
Layer 1 β Research CLAUDE.md Template
Instead of writing a CLAUDE.md from scratch for each project, I created a reusable template at ~/.claude/templates/research-CLAUDE.md with pre-structured sections:
# Research Project:
> PI: ... | Started: ... | Status: ...
## Project Goal β 1 sentence
## Research Questions β numbered RQs
## Key Hypotheses β with predictions
## Decisions Log β date | decision | rationale
## Data & Artifacts β what | location | format
## Code Standards β language, build, test commands
## Current Progress β done / in progress / next
## Open Questions β things that need answering
## Key References β with why-it-matters annotations
## Persona Activation β which commands to use
## Parallel Agent Workflow β copy-paste recipes
For my cell morphing project, the CLAUDE.md includes solver-specific instructions: βalways use pcsat_tbq_ar.json configβ, βnever use integer division in goal constraintsβ, βrun non-vacuity check after every satβ. These rules get enforced automatically across every session.
Layer 2 β Five Persona Commands
I created these as global slash commands (in ~/.claude/commands/), so they work in any project:
| Command | What It Actually Contains |
|---|---|
/reviewer2 |
Full review checklist (correctness, soundness, completeness for code; claims vs evidence, methodology, related work for papers). Outputs structured report with REJECT/MAJOR/MINOR/ACCEPT verdict. |
/editor |
4-pass editing pipeline: Structure β Clarity β Precision β Polish. Knows CS/PL/security writing conventions. Shows beforeβafter quality score. |
/stats-checker |
Validates assumptions (normality, independence, homogeneity), checks test selection, flags missing effect sizes. CS-specific: benchmark fairness, warmup runs, timing variance. |
/lit-scout |
Searches multiple databases, organizes into must-cite / directly related / methodologically related / recent. Drafts a related work paragraph. Prioritizes PL/security venues (PLDI, POPL, CAV, S&P). |
/viz-designer |
Decision tree for choosing visualization type. Defaults to colorblind-safe palettes, vector formats for papers. Generates LaTeX/TikZ for publication figures. |
Each command is ~80 lines of specific instructions β not a vague βbe a reviewerβ, but a checklist-driven workflow that produces consistent, structured output.
Layer 3 β Parallel Agent Recipes
Built into the CLAUDE.md as copy-paste workflows:
## Pre-submission Review Sprint
Agent 1: /reviewer2 paper/ β hostile review
Agent 2: /stats-checker paper/ β validate methods
Agent 3: /editor paper/intro.tex β polish introduction
## Literature + Writing Sprint
Agent 1: /lit-scout "relational cost verification CHC"
Agent 2: /editor paper/related.tex
Agent 3: /viz-designer "benchmark comparison chart"
The Bootstrap Command
The key insight: make setup frictionless. I created /research-setup β a command you run inside any existing project folder:
cd ~/my-research-project
# Then in Claude Code:
/research-setup
It scans the directory, infers the project type from existing files (.clp β formal methods, .py β Python, .tex β paper), and adds only whatβs missing. Never overwrites, never creates unwanted directories, asks before changing anything.
What Changed in Practice
| Before | After |
|---|---|
| Re-explain project context every session | CLAUDE.md remembers everything |
| Generic βreview my codeβ prompts | /reviewer2 with structured checklist |
| Manual literature searches | /lit-scout with database integration |
| One task at a time | Three agents running in parallel |
| 30 min setting up each new project | /research-setup in 2 minutes |
The biggest win isnβt any single persona β itβs the compound effect. When CLAUDE.md remembers your project state, personas know how to review your specific domain, and you can run three agents while thinking about the next experiment β youβre operating at a fundamentally different pace.
How LearnAI Team Could Use This
- Adapt this model into a reusable research and documentation infrastructure: CLAUDE.md files for project memory, persona commands for review/editing/literature work.
- Run parallel sessions for audits, synthesis, and publishing workflows.
- Bootstrap new research projects with repeatable templates and commands.
Real-World Use Cases
- Maintain project-specific CLAUDE.md files for LearnAI research, docs, and course repos.
- Use reviewer/editor/lit-scout personas to improve articles before publication.
- Run parallel Claude Code sessions for code audit, source verification, and summary drafting.
Getting Started
- Create a CLAUDE.md for your research project β or use
/research-setupto bootstrap one from a template - Build 2-3 personas β start with βReviewer #2β and βEditorβ as slash commands in
~/.claude/commands/ - Try one parallel session β run a literature search agent while you work on analysis
- Iterate β update CLAUDE.md as your project evolves
- Make it reusable β create a
/research-setupcommand so every new project gets the infrastructure automatically
Resources: ClaudeCodeTools Presentation (PDF)