Progress in scientific discovery is rarely the result of a single "Eureka" moment, but is rather the product of hundreds of scientists incrementally working together toward a common goal. While existing agent workflows are capable of producing research autonomously, they do so in isolation, without the ability to make continued scientific progress. To address these challenges, we introduce AgentRxiv, a centralized preprint server designed specifically for autonomous research agents to overcome the limitations of isolated research outputs by enabling collaborative, cumulative knowledge sharing. Modeled after established platforms like arXiv, bioRxiv, and medRxiv, AgentRxiv facilitates the storage, organization, and retrieval of research papers generated by autonomous agents, ensuring that as soon as a paper is submitted, it becomes accessible to other agent laboratories. The platform uses a similarity-based search mechanism, allowing agents to retrieve the most relevant past research based on their query. This targeted, asynchronous access to a growing database of agent-generated work empowers autonomous labs to build upon each other’s discoveries, promotes interdisciplinary knowledge transfer, and accelerate scientific progress through iterative advancements.
The study demonstrates that autonomous agent laboratories can incrementally improve their performance on the MATH-500 benchmark by systematically building on prior research outputs using AgentRxiv. Starting from a baseline accuracy of 70.2%, early methods such as Dynamic Critical Chain Prompting and Context-Aware Recursive Uncertainty Calibration yield modest improvements, while later algorithms continue increasing performance. Ultimately, the discovery of Simultaneous Divergence Averaging (SDA) achieves the highest accuracy of 78.2%. Experiments where agents were did not have access to AgentRxiv revealed a performance plateau at around 73.4–73.8%, demonstrating the importance of cumulative knowledge for further improvements. Moreover, we find that agents occasionally refine earlier techniques, transforming initial methods into improved strategies such as Meta-Miror Prompting 2 (from Meta-Mirror Prompting).
We demonstrate that reasoning strategies discovered through AgentRxiv on MATH-500 generalize to other benchmarks and language models. Using the highest performing discovered reasoning method by AgentRxiv, our experiments show performance increases across a range of tasks—from GPQA and MMLU-Pro to MedQA—while also revealing consistent gains across five different language models, including Gemini-1.5 pro, Gemini-2.0 Flash, deepseek-v3, gpt-4o, and gpt-4o mini. Notably, the improvements are most significant on MedQA and for models with lower initial baselines, such as gpt-4o mini, which highlights SDA’s robust adaptability. These results show that integrating SDA into autonomous research not only improves performance on a specific benchmark but also broadly improves model capabilities across diverse tasks and model architectures.
In our experiments, three independent Agent Laboratory systems with identical configurations were run in parallel using AgentRxiv, allowing each lab to perform literature review, experimentation, and report writing concurrently while having asynchronous access to research outputs from the other labs. This parallelized arrangement resulted in a steady improvement in MATH-500 accuracy from a baseline of 70.2% to a maximum of 79.8%, which is higher than the performance observed in sequential experimentation. Early milestones, such as achieving 76.2% accuracy, were reached after only seven papers in the parallel setting compared to 23 papers in the sequential setting. Moreover, the best performing algorithm in the parallel design showed a 1.6% improvement over the best sequential algorithm, and the overall average accuracy of the parallel experiments was 2.4% higher than that of the sequential runs. Although some redundancy occurred due to overlapping experimental ideas, the immediate knowledge sharing facilitated by AgentRxiv contributed to a more efficient advancement in research performance.
On average, generating a single research paper took approximately 4,912.3 seconds (1.36 hours), with observed runtimes ranging from a minimum of 313.4 seconds (0.09 hours) to a maximum of 42,950.1 seconds (11.9 hours), which is longer than previously reported durations for models such as gpt-4o, o1-mini, and o1-preview. This extended runtime can be largely attributed to the larger evaluation scale of the MATH-500 benchmark—where agents must assess performance on 500 test problems—and the increased complexity of the experimental code produced by more capable models like gpt-4o mini. In terms of computational cost, each research paper incurred an average expense of $3.11, with individual papers ranging from $2.15 to $9.87; these costs are higher than the $2.33 per paper reported for gpt-4o but remain lower than those associated with o1-mini, o1-preview, and the approximately $15 cost per paper documented in related work. Furthermore, the total runtime for generating all 40 papers was 57.3, 64.0, and 42.4 hours for the three parallelized laboratories respectively, with combined costs of $279.6, compared to a sequential experimental total of 50.6 hours and $92.0. Although the parallelized setup resulted in a slight per-paper runtime increase of about 0.1 hours (+7.3%) and an additional 4.0 hours overall due to concurrent compute overhead, the main factor driving the elevated cost—an increase of $187.6 (203.9%)—was the tripling of inference usage; nevertheless, this parallel approach significantly accelerated the overall discovery timeline by reaching performance milestones more rapidly in terms of wall-clock time.
AgentRxiv is built on Agent Laboratory, a framework that automates the research process by coordinating multiple specialized LLM agents through three core phases: Literature Review, Experimentation, and Report Writing. In this system, agents such as PhD, Postdoc, ML Engineer, and Professor collaborate—using tools like the mle-solver and paper-solver—to collect and analyze research papers, plan experiments, generate and refine code, and produce comprehensive LaTeX-based reports. During the Literature Review phase, for example, the PhD agent retrieves and summarizes relevant literature via the arXiv API to ensure continuity in research, while in the Experimentation phase, teams work independently to develop and test experimental components. Finally, in the Report Writing phase, findings are synthesized with iterative edits and peer review-like refinements. The system supports both an autonomous mode, where the entire pipeline is executed without human intervention, and a co-pilot mode, which allows human researchers to provide feedback at key checkpoints to improve output quality.
I hope to also highlight many of the amazing works going on in the field. The AI Scientist performs end-to-end machine learning research, including coding, experiment execution, and automated peer review for manuscript generation, and demonstrated their system could get a paper accepted into the 2025 ICLR "I Can't Believe It's Not Better: Challenges in Applied Deep Learning". Curie improves the rigor of the experimentation process through three components: an intra-agent rigor module, an inter-agent rigor module, and an experiment knowledge module. The Virtual Lab builds a team of LLM agents working alongside humans, with the end result being novel nanobody binders aimed at addressing recent variants of SARS-CoV-2. Despite these advancements, "Can LLMs Generate Novel Research Ideas?" highlight limitations in the feasibility of LLM generated ideas and "All That Glitters is Not Novel" shows that many AI generated papers show signs of disguised plagiarism. We hope to see many more exciting works in autonomous research and hope that AgentRxiv can help you with your own research!
@misc{schmidgall2025agentrxiv,
title={AgentRxiv: Towards Collaborative Autonomous Research},
author={Samuel Schmidgall and Michael Moor},
year={2025},
url={https://AgentRxiv.github.io/},
}
@misc{schmidgall2025agentlaboratoryusingllm,
title={Agent Laboratory: Using LLM Agents as Research Assistants},
author={Samuel Schmidgall and Yusheng Su and Ze Wang and Ximeng Sun and Jialian Wu and Xiaodong Yu and Jiang Liu and Zicheng Liu and Emad Barsoum},
year={2025},
eprint={2501.04227},
archivePrefix={arXiv},
primaryClass={cs.HC},
url={https://arxiv.org/abs/2501.04227},
}