aiming-lab / AutoResearchClaw
Fully autonomous research from idea to paper. Chat an Idea. Get a Paper. Fully Autonomous. π¦
AI Architecture Analysis
This repository is indexed by RepoMind. By analyzing aiming-lab/AutoResearchClaw in our AI interface, you can instantly generate complete architecture diagrams, visualize control flows, and perform automated security audits across the entire codebase.
Our Agentic Context Augmented Generation (Agentic CAG) engine loads full source files into context on-demand, avoiding the fragmentation of traditional RAG systems. Ask questions about the architecture, dependencies, or specific features to see it in action.
Repository Overview (README excerpt)
Crawler viewChat an Idea. Get a Paper. Fully Autonomous. Just chat with OpenClaw : "Research X" β done. π¨π³ δΈζ Β· π―π΅ ζ₯ζ¬θͺ Β· π°π· νκ΅μ΄ Β· π«π· FranΓ§ais Β· π©πͺ Deutsch Β· πͺπΈ EspaΓ±ol Β· π§π· PortuguΓͺs Β· π·πΊ Π ΡΡΡΠΊΠΈΠΉ Β· πΈπ¦ Ψ§ΩΨΉΨ±Ψ¨ΩΨ© π Integration Guide Β· π¬ Discord Community --- > **π§ͺ We're looking for testers!** Try the pipeline with your own research idea β from any field β and tell us what you think. Your feedback directly shapes the next version. **β Testing Guide** | **β δΈζζ΅θ―ζε** --- π₯ News β’ **[03/17/2026]** v0.3.0 β **MetaClaw Integration** β AutoResearchClaw now supports MetaClaw cross-run learning: pipeline failures β structured lessons β reusable skills, injected into all 23 stages. **+18.3%** robustness in controlled experiments. Opt-in ( ), fully backward-compatible. See Integration Guide. β’ **[03/16/2026]** v0.2.0 β Three multi-agent subsystems (CodeAgent, BenchmarkAgent, FigureAgent), hardened Docker sandbox with network-policy-aware execution, 4-round paper quality audit (AI-slop detection, 7-dim review scoring, NeurIPS checklist), and 15+ bug fixes from production runs. β’ **[03/15/2026]** v0.1.0 β We release AutoResearchClaw: a fully autonomous 23-stage research pipeline that turns a single research idea into a conference-ready paper. No human intervention required. --- β‘ One Command. One Paper. --- π€ What Is This? **You think it. AutoResearchClaw writes it.** Drop a research topic β get back a full academic paper with real literature from OpenAlex, Semantic Scholar & arXiv, hardware-aware sandbox experiments (GPU/MPS/CPU auto-detected), statistical analysis, multi-agent peer review, and conference-ready LaTeX targeting NeurIPS/ICML/ICLR. No babysitting. No copy-pasting. No hallucinated references. π paper_draft.md Full academic paper (Introduction, Related Work, Method, Experiments, Results, Conclusion) π paper.tex Conference-ready LaTeX (NeurIPS / ICLR / ICML templates) π references.bib Real BibTeX references from OpenAlex, Semantic Scholar and arXiv β auto-pruned to match inline citations π verification_report.json 4-layer citation integrity + relevance verification (arXiv, CrossRef, DataCite, LLM) π§ͺ experiment runs/ Generated code + sandbox results + structured JSON metrics π charts/ Auto-generated condition comparison charts with error bars and confidence intervals π reviews.md Multi-agent peer review with methodology-evidence consistency checks 𧬠evolution/ Self-learning lessons extracted from each run π¦ deliverables/ All final outputs in one folder β compile-ready for Overleaf The pipeline runs **end-to-end without human intervention**. When experiments fail, it self-heals. When hypotheses don't hold, it pivots. When citations are fake, it kills them. --- π Quick Start Output β β compile-ready LaTeX, BibTeX, experiment code, charts. π Minimum required config --- π§ What Makes It Different | Capability | How It Works | |-----------|-------------| | **π PIVOT / REFINE Loop** | Stage 15 autonomously decides: PROCEED, REFINE (tweak params), or PIVOT (new direction). Artifacts auto-versioned. | | **π€ Multi-Agent Debate** | Hypothesis generation, result analysis, and peer review each use structured multi-perspective debate. | | **𧬠Self-Learning** | Lessons extracted per run (decision rationale, runtime warnings, metric anomalies) with 30-day time-decay. Future runs learn from past mistakes. | | **π Knowledge Base** | Every run builds structured KB across 6 categories (decisions, experiments, findings, literature, questions, reviews). | | **π‘οΈ Sentinel Watchdog** | Background quality monitor: NaN/Inf detection, paper-evidence consistency, citation relevance scoring, anti-fabrication guard. | --- π¦ OpenClaw Integration **AutoResearchClaw is an OpenClaw-compatible service.** Install it in OpenClaw and launch autonomous research with a single message β or use it standalone via CLI, Claude Code, or any AI coding assistant. π Use with OpenClaw (Recommended) If you already use OpenClaw as your AI assistant: **That's it.** OpenClaw handles , , config setup, and pipeline execution automatically. You just chat. π‘ What happens under the hood β’ OpenClaw reads β learns the research orchestrator role β’ OpenClaw reads β understands installation and pipeline structure β’ OpenClaw copies β β’ Asks for your LLM API key (or uses your environment variable) β’ Runs + β’ Returns the paper, LaTeX, experiments, and citations π OpenClaw Bridge (Advanced) For deeper integration, AutoResearchClaw includes a **bridge adapter system** with 6 optional capabilities: Each flag activates a typed adapter protocol. When OpenClaw provides these capabilities, the adapters consume them without code changes. See for full details. ACP (Agent Client Protocol) AutoResearchClaw can use **any ACP-compatible coding agent** as its LLM backend β no API keys required. The agent communicates via acpx, maintaining a single persistent session across all 23 pipeline stages. | Agent | Command | Notes | |-------|---------|-------| | Claude Code | | Anthropic | | Codex CLI | | OpenAI | | Gemini CLI | | Google | | OpenCode | | SST | | Kimi CLI | | Moonshot | π οΈ Other Ways to Run | Method | How | |--------|-----| | **Standalone CLI** | | | **Python API** | | | **Claude Code** | Reads β just say *"Run research on [topic]"* | | **OpenCode** | Reads β same natural language interface | | **Any AI CLI** | Provide as context β agent auto-bootstraps | --- π¬ Pipeline: 23 Stages, 8 Phases > **Gate stages** (5, 9, 20) pause for human approval or auto-approve with . On rejection, the pipeline rolls back. > **Decision loops**: Stage 15 can trigger REFINE (β Stage 13) or PIVOT (β Stage 8), with automatic artifact versioning. π What Each Phase Does | Phase | What Happens | |-------|-------------| | **A: Scoping** | LLM decomposes the topic into a structured problem tree with research questions | | **A+: Hardware** | Auto-detects GPU (NVIDIA CUDA / Apple MPS / CPU-only),β¦