cd ../
πŸŽ“ Research & Academia
Requires OpenClaw v2026.2+|RAG + Zotero

How OpenClaw Helped Me Survive My PhD Literature Review

By u/BioPhDSurvivorβ€’February 25, 2026β€’ 312 comments

I'm a 3rd-year PhD student in computational biology. My literature review required analyzing 400+ papers across 15 journals. Manual note-taking in Zotero was killing my wrists and my sanity. OpenClaw turned a 4-month nightmare into a 3-week sprint.

The Paper Mountain

400 papers. 15 journals. Finding connections between methods, identifying gaps, tracking citation networks. Manual note-taking in Zotero was killing my wrists and my sanity.

400+
Papers to Review
15
Journals
4 months
Estimated Time
~12
Manual Notes/Day

Architecture Overview

OpenClaw runs on my lab's Ubuntu workstation (32GB RAM, RTX 4070). It connects to my Zotero library via the Zotero API, indexes all PDFs locally using a RAG pipeline with sentence-transformers, and uses Mixtral-8x7B via Ollama for synthesis and analysis.

β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”    API      β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚   Zotero    │◄────────────│   OpenClaw    β”‚
β”‚   Library   │────────────►│   Agent      β”‚
β”‚  (400+ PDFs)β”‚             β””β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”˜
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜                    β”‚
                     β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
                     β–Ό             β–Ό             β–Ό
              β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β” β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
              β”‚ ChromaDB  β”‚ β”‚  Ollama  β”‚ β”‚ Obsidian β”‚
              β”‚ Vector DB β”‚ β”‚ Mixtral  β”‚ β”‚  Notes   β”‚
              β”‚ (RAG)     β”‚ β”‚ 8x7B    β”‚ β”‚  Export  β”‚
              β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜

OpenClaw Configuration

IDENTITY.md
# IDENTITY.md for Research Assistant

You are a research assistant for a computational biology PhD student.
Your role is to help with literature review, paper analysis, and synthesis.

## Core Capabilities
1. Index and search 400+ papers using semantic similarity (ChromaDB)
2. Generate structured comparison tables across papers
3. Identify contradictions and gaps in the literature
4. Draft literature review paragraphs with proper citations
5. Track citation networks and find influential papers

## Search Behavior
- Use semantic search, not keyword matching
- Always include: paper title, authors, year, journal, DOI
- Rank results by relevance AND recency
- Flag retracted papers or significant erratum

## Synthesis Rules
- Never fabricate citations β€” only reference indexed papers
- Always distinguish between "paper claims" vs "paper proves"
- Flag statistical issues: small sample size (<30), p-hacking signs
- When papers contradict, present BOTH sides neutrally
- Use APA 7th edition citation format

## Writing Assistance
- Match the academic voice of the field (formal, passive voice ok)
- Every claim must have a citation
- Highlight when a gap in the literature exists
- Never generate "review paper" style writing β€” always specific claims
setup.sh
# Setup: RAG Pipeline for Paper Indexing

# 1. Install dependencies
pip install chromadb sentence-transformers pymupdf

# 2. Index all Zotero PDFs
python index_papers.py --zotero-dir ~/Zotero/storage \
  --model all-MiniLM-L6-v2 \
  --chunk-size 512 \
  --overlap 64 \
  --db-path ./chroma_papers

# 3. Configure OpenClaw environment
export OPENCLAW_MODEL=ollama:mixtral:8x7b
export CHROMA_DB_PATH=./chroma_papers
export ZOTERO_API_KEY=your_api_key_here
export ZOTERO_USER_ID=your_user_id
export OBSIDIAN_VAULT=~/Documents/PhD-Notes

# Index stats after completion:
# Papers indexed: 412
# Total chunks: 47,832
# Embedding model: all-MiniLM-L6-v2 (384-dim)
# Index size: 2.3 GB
# Indexing time: 23 minutes (RTX 4070)

1. Natural Language Paper Search

Instead of boolean database searches, I ask questions in plain English. OpenClaw searches my indexed library using semantic similarity and returns ranked results with relevant excerpts.

Query: "Find papers that use transformer architectures for 
protein folding AND cite AlphaFold 2 but NOT Rosetta"

Results (0.3s):
β”Œβ”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚ #  β”‚ Paper                                        β”‚ Year β”‚ Relevance β”‚
β”œβ”€β”€β”€β”€β”Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”Όβ”€β”€β”€β”€β”€β”€β”Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€
β”‚ 1  β”‚ Lin et al. "ESMFold: Language models enable   β”‚ 2023 β”‚ 0.94      β”‚
β”‚    β”‚ zero-shot prediction of protein structure"    β”‚      β”‚           β”‚
β”‚ 2  β”‚ Wu et al. "High-resolution de novo structure  β”‚ 2024 β”‚ 0.91      β”‚
β”‚    β”‚ prediction from primary sequence"             β”‚      β”‚           β”‚
β”‚ 3  β”‚ Ahdritz et al. "OpenFold: Lessons learned"    β”‚ 2024 β”‚ 0.89      β”‚
β”‚ 4  β”‚ Baek et al. "Efficient backbone generation"   β”‚ 2023 β”‚ 0.86      β”‚
β”‚ 5  β”‚ Rives et al. "Biological structure from       β”‚ 2023 β”‚ 0.84      β”‚
β”‚    β”‚ scaling unsupervised learning"                β”‚      β”‚           β”‚
β””β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜

Key excerpt from #1 (p.4):
"Unlike Rosetta, which relies on physics-based energy functions,
ESMFold leverages a transformer architecture trained on 65M 
protein sequences, achieving competitive accuracy with AF2..."

2. Automated Paper Synthesis

OpenClaw reads batches of papers and generates structured comparison tables. It identifies methodological differences, contradictions, and gaps β€” saving weeks of manual comparison.

Command: "Compare methods in papers 1-5 regarding protein 
structure prediction accuracy"

Generated Comparison Table:
β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”¬β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚ Paper     β”‚ Method     β”‚ Dataset    β”‚ GDT-TS    β”‚ Key Limitat. β”‚
β”œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”Όβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€
β”‚ Lin 2023  β”‚ ESMFold    β”‚ CASP15     β”‚ 0.71      β”‚ No MSA input β”‚
β”‚ Wu 2024   β”‚ RFdiffuse  β”‚ CATH 4.3   β”‚ 0.83      β”‚ High compute β”‚
β”‚ Ahdritz   β”‚ OpenFold   β”‚ CASP15     β”‚ 0.79      β”‚ Training costβ”‚
β”‚ Baek 2023 β”‚ RoseTTAFoldβ”‚ PDB 2022   β”‚ 0.76      β”‚ Memory limit β”‚
β”‚ Rives 2023β”‚ ESM-1v     β”‚ ProteinNet β”‚ 0.68      β”‚ Single seq.  β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”΄β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜

⚠️ CONTRADICTION DETECTED:
Lin 2023 reports GDT-TS 0.71 for ESMFold on CASP15,
but Rives 2023 reports 0.68 for a similar model (ESM-1v).
Possible explanation: different evaluation splits.
β†’ Flagged for manual verification.

πŸ“Š GAP IDENTIFIED:
No paper in this set evaluates on membrane proteins specifically.
All benchmarks use soluble protein datasets. This is a potential
thesis contribution.

3. Drafting with Proper Citations

For each subsection, I give OpenClaw the relevant papers and my thesis statement. It drafts paragraphs with proper APA citations. I edit ~40% but the structure and citations are always solid.

Input: "Draft a paragraph about transformer-based approaches 
to protein folding, using papers 1, 3, and 5"

Draft Output:
"Recent advances in transformer architectures have demonstrated 
remarkable potential for protein structure prediction. Lin et al. 
(2023) introduced ESMFold, which leverages a 15-billion parameter 
language model trained on 65 million protein sequences, achieving 
competitive accuracy with AlphaFold 2 while requiring only a 
single sequence as input β€” eliminating the need for multiple 
sequence alignments (MSAs). This approach was further validated 
by Ahdritz et al. (2024), who reproduced and extended the 
methodology through their OpenFold framework, demonstrating that 
training transparency and reproducibility are achievable without 
sacrificing predictive performance (GDT-TS: 0.79 on CASP15). 
However, Rives et al. (2023) noted that single-sequence models 
still underperform MSA-based methods for proteins with fewer 
homologs, suggesting that the transformer's self-attention 
mechanism may not fully capture co-evolutionary information 
present in aligned sequences."

Citations auto-generated:
[1] Lin, Z., et al. (2023). Science, 379(6637), 1123-1130.
[3] Ahdritz, G., et al. (2024). Nature Methods, 21(1), 45-52.
[5] Rives, A., et al. (2023). PNAS, 120(15), e2016239120.

4. Citation Network Mapping

OpenClaw tracks which papers cite which, identifies highly-cited foundational works, and finds unexpected connections between sub-fields.

Command: "Map the citation network for AlphaFold 2 derivatives"

Citation Graph (top 10 by in-degree):
Jumper 2021 (AlphaFold 2) ───┬── 847 citations in my corpus
                              β”œβ”€β”€ Lin 2023 (ESMFold)
                              β”œβ”€β”€ Baek 2023 (RoseTTAFold)
                              β”œβ”€β”€ Ahdritz 2024 (OpenFold)
                              └── Wu 2024 (RFdiffusion)

πŸ”— UNEXPECTED CONNECTION:
Paper by Chen 2024 (drug discovery) cites both AlphaFold 2
AND a 2019 paper by Krishnamurthy (computational ecology).
The ecology paper uses protein folding techniques for 
environmental DNA analysis. This cross-pollination was not 
obvious from keyword searches.

β†’ Added to "Interdisciplinary Applications" section of thesis

The Outcome

Literature review completed in 3 weeks β€” here's the quantitative comparison:

MetricManual ProcessWith OpenClawChange
Time to complete~4 months3 weeks↓ 81%
Papers analyzed~150 (gave up)412↑ 175%
Contradictions found27↑ 250%
Cross-field connections03New finding
Draft paragraphs/day2-315-20↑ 600%
Citation accuracy94%99.5%↑ 5.5%
"My advisor asked how I found a 2019 paper from an obscure ecology journal that perfectly contradicted our hypothesis. I just said 'I did a thorough search.' I didn't mention my AI intern." β€” u/BioPhDSurvivor

Cost Analysis

ItemCostNotes
Lab workstation$0Already in lab (shared)
Ollama + Mixtral-8x7B$0Self-hosted, RTX 4070
ChromaDB$0Open source, local
Zotero (free tier)$0300MB cloud, unlimited local
Sentence-transformers$0Open source model
Total$0/movs $200+/mo for commercial tools

Zero additional cost β€” used existing lab hardware. Equivalent commercial tools (Elicit, Consensus, Semantic Scholar Premium) would cost $200+/month.

Academic Integrity & Privacy

All papers processed locally β€” no upload to cloud APIs
OpenClaw NEVER generates new data or fabricates results
Every claim in drafts has a verifiable citation link
I manually verify every citation against the source paper
Draft text is starting material, not final submission
Advisor aware of tool usage β€” department allows AI writing assistants

⚠️ AI-assisted writing requires disclosure per most university policies. Check your department's guidelines. The output is a DRAFT β€” you must verify and edit substantially.

Frequently Asked Questions

Q1. Doesn't this count as cheating?

My university's policy allows AI writing assistants for drafting and research. The key: I edit ~40% of the text, verify every citation, and my own analysis/interpretation is original. OpenClaw is a research tool, not a ghostwriter. I disclosed its use to my advisor.

Q2. What model do you recommend for academic work?

Mixtral-8x7B is the sweet spot: good reasoning for synthesis tasks, runs on a 12GB+ GPU. Llama-3-8B struggles with nuanced comparison. GPT-4 is better but sends your unpublished research to OpenAI's servers β€” unacceptable for pre-publication work.

Q3. How accurate are the generated citations?

99.5% accurate because it ONLY cites papers in my indexed library (ChromaDB). It can't make up papers that don't exist in the index. The 0.5% error comes from occasionally attributing a claim to the wrong section of a paper.

Q4. Can it handle non-English papers?

Partially. sentence-transformers works well for abstracts (usually English). For full-text Chinese/Japanese papers, I first run them through a translation pipeline. Plan to switch to multilingual-e5-large for native multi-language support.

Q5. What about figures and tables in papers?

Currently text-only. PDFs are extracted using PyMuPDF which doesn't parse figures. For table extraction, I'm testing Nougat (Meta's academic OCR model) but it's not integrated yet.

Lessons Learned

Chunk size matters enormously

512-token chunks with 64-token overlap was the sweet spot. Too small (128) loses context. Too large (1024) dilutes relevance scores. This single tweak improved search accuracy by 25%.

Never trust AI citations blindly

Even with a constrained index, I spot-check every citation. In 412 papers, I found 2 cases where OpenClaw attributed a finding to the wrong paper. Always verify.

Use it for search and structure, not analysis

OpenClaw excels at finding relevant papers and organizing them. It's mediocre at deep analysis. My best results came from using it to surface papers I'd never have found, then doing my own analysis.

Export notes to Obsidian for permanence

I export all synthesis tables and connection maps to Obsidian Markdown. This way, even if the OpenClaw setup changes, my research notes persist in a portable format.