> research-pipeline
Full research pipeline: Workflow 1 (idea discovery) → implementation → Workflow 2 (auto review loop). Goes from a broad research direction all the way to a submission-ready paper. Use when user says "全流程", "full pipeline", "从找idea到投稿", "end-to-end research", or wants the complete autonomous research lifecycle.
curl "https://skillshub.wtf/wanshuiyin/Auto-claude-code-research-in-sleep/research-pipeline?format=md"Full Research Pipeline: Idea → Experiments → Submission
End-to-end autonomous research workflow for: $ARGUMENTS
Constants
- AUTO_PROCEED = true — When
true, Gate 1 auto-selects the top-ranked idea (highest pilot signal + novelty confirmed) and continues to implementation. Whenfalse, always waits for explicit user confirmation before proceeding. - ARXIV_DOWNLOAD = false — When
true,/research-litdownloads the top relevant arXiv PDFs during literature survey. Whenfalse(default), only fetches metadata via arXiv API. Passed through to/idea-discovery→/research-lit. - HUMAN_CHECKPOINT = false — When
true, the auto-review loops (Stage 4) pause after each round's review to let you see the score and provide custom modification instructions before fixes are implemented. Whenfalse(default), loops run fully autonomously. Passed through to/auto-review-loop.
💡 Override via argument, e.g.,
/research-pipeline "topic" — AUTO_PROCEED: false, human checkpoint: true.
Overview
This skill chains the entire research lifecycle into a single pipeline:
/idea-discovery → implement → /run-experiment → /auto-review-loop → submission-ready
├── Workflow 1 ──┤ ├────────── Workflow 2 ──────────────┤
It orchestrates two major workflows plus the implementation bridge between them.
Pipeline
Stage 1: Idea Discovery (Workflow 1)
Invoke the idea discovery pipeline:
/idea-discovery "$ARGUMENTS"
This internally runs: /research-lit → /idea-creator → /novelty-check → /research-review
Output: IDEA_REPORT.md with ranked, validated, pilot-tested ideas.
🚦 Gate 1 — Human Checkpoint:
After IDEA_REPORT.md is generated, pause and present the top ideas to the user:
📋 Idea Discovery complete. Top ideas:
1. [Idea 1 title] — Pilot: POSITIVE (+X%), Novelty: CONFIRMED
2. [Idea 2 title] — Pilot: WEAK POSITIVE (+Y%), Novelty: CONFIRMED
3. [Idea 3 title] — Pilot: NEGATIVE, eliminated
Recommended: Idea 1. Shall I proceed with implementation?
If AUTO_PROCEED=false: Wait for user confirmation before continuing. The user may:
- Approve an idea → proceed to Stage 2.
- Pick a different idea → proceed with their choice.
- Request changes (e.g., "combine Idea 1 and 3", "focus more on X") → update the idea prompt with user feedback, re-run
/idea-discoverywith refined constraints, and present again. - Reject all ideas → collect feedback on what's missing, re-run Stage 1 with adjusted research direction. Repeat until the user commits to an idea.
- Stop here → save current state to
IDEA_REPORT.mdfor future reference.
If AUTO_PROCEED=true: Present the top ideas, wait 10 seconds for user input. If no response, auto-select the #1 ranked idea (highest pilot signal + novelty confirmed) and proceed to Stage 2. Log: "AUTO_PROCEED: selected Idea 1 — [title]".
⚠️ This gate waits for user confirmation when AUTO_PROCEED=false. When
true, it auto-selects the top idea after presenting results. The rest of the pipeline (Stages 2-4) is expensive (GPU time + multiple review rounds), so setAUTO_PROCEED=falseif you want to manually choose which idea to pursue.
Stage 2: Implementation
Once the user confirms which idea to pursue:
-
Read the idea details from
IDEA_REPORT.md(hypothesis, experimental design, pilot code) -
Implement the full experiment:
- Extend pilot code to full scale (multi-seed, full dataset, proper baselines)
- Add proper evaluation metrics and logging (wandb if configured)
- Write clean, reproducible experiment scripts
- Follow existing codebase conventions
-
Code review: Before deploying, do a self-review:
- Are all hyperparameters configurable via argparse?
- Is the random seed fixed and controllable?
- Are results saved to JSON/CSV for later analysis?
- Is there proper logging for debugging?
Stage 3: Deploy Experiments (Workflow 2 — Part 1)
Deploy the full-scale experiments:
/run-experiment [experiment command]
What this does:
- Check GPU availability on configured servers
- Sync code to remote server
- Launch experiments in screen sessions with proper CUDA_VISIBLE_DEVICES
- Verify experiments started successfully
Monitor progress:
/monitor-experiment [server]
Wait for experiments to complete. Collect results.
Stage 4: Auto Review Loop (Workflow 2 — Part 2)
Once initial results are in, start the autonomous improvement loop:
/auto-review-loop "$ARGUMENTS — [chosen idea title]"
What this does (up to 4 rounds):
- GPT-5.4 xhigh reviews the work (score, weaknesses, minimum fixes)
- Claude Code implements fixes (code changes, new experiments, reframing)
- Deploy fixes, collect new results
- Re-review → repeat until score ≥ 6/10 or 4 rounds reached
Output: AUTO_REVIEW.md with full review history and final assessment.
Stage 5: Final Summary
After the auto-review loop completes, write a final status report:
# Research Pipeline Report
**Direction**: $ARGUMENTS
**Chosen Idea**: [title]
**Date**: [start] → [end]
**Pipeline**: idea-discovery → implement → run-experiment → auto-review-loop
## Journey Summary
- Ideas generated: X → filtered to Y → piloted Z → chose 1
- Implementation: [brief description of what was built]
- Experiments: [number of GPU experiments, total compute time]
- Review rounds: N/4, final score: X/10
## Final Status
- [ ] Ready for submission / [ ] Needs manual follow-up
## Remaining TODOs (if any)
- [items flagged by reviewer that weren't addressed]
## Files Changed
- [list of key files created/modified]
Key Rules
-
Large file handling: If the Write tool fails due to file size, immediately retry using Bash (
cat << 'EOF' > file) to write in chunks. Do NOT ask the user for permission — just do it silently. -
Human checkpoint after Stage 1 is controlled by AUTO_PROCEED. When
false, do not proceed without user confirmation. Whentrue, auto-select the top idea after presenting results. -
Stages 2-4 can run autonomously once the user confirms the idea. This is the "sleep and wake up to results" part.
-
If Stage 4 ends at round 4 without positive assessment, stop and report remaining issues. Do not loop forever.
-
Budget awareness: Track total GPU-hours across the pipeline. Flag if approaching user-defined limits.
-
Documentation: Every stage updates its own output file. The full history should be self-contained.
-
Fail gracefully: If any stage fails (no good ideas, experiments crash, review loop stuck), report clearly and suggest alternatives rather than forcing forward.
Typical Timeline
| Stage | Duration | Can sleep? |
|---|---|---|
| 1. Idea Discovery | 30-60 min | Yes if AUTO_PROCEED=true |
| 2. Implementation | 15-60 min | Yes (autonomous after Gate 1) |
| 3. Deploy | 5 min + experiment time | Yes ✅ |
| 4. Auto Review | 1-4 hours (depends on experiments) | Yes ✅ |
Sweet spot: Run Stage 1-2 in the evening, launch Stage 3-4 before bed, wake up to a reviewed paper.
> related_skills --same-repo
> run-experiment
Deploy and run ML experiments on local or remote GPU servers. Use when user says "run experiment", "deploy to server", "跑实验", or needs to launch training jobs.
> research-review
Get a deep critical review of research from GPT via Codex MCP. Use when user says "review my research", "help me review", "get external review", or wants critical feedback on research ideas, papers, or experimental results.
> research-refine
Turn a vague research direction into a problem-anchored, elegant, frontier-aware, implementation-oriented method plan via iterative GPT-5.4 review. Use when the user says "refine my approach", "帮我细化方案", "decompose this problem", "打磨idea", "refine research plan", "细化研究方案", or wants a concrete research method that stays simple, focused, and top-venue ready instead of a vague or overbuilt idea.
> research-refine-pipeline
Run an end-to-end workflow that chains `research-refine` and `experiment-plan`. Use when the user wants a one-shot pipeline from vague research direction to focused final proposal plus detailed experiment roadmap, or asks to "串起来", build a pipeline, do it end-to-end, or generate both the method and experiment plan together.