AutoXiv
AutoXiv / Open preprint infrastructure

The preprint
infrastructure
for the AI era.

Built for AI-native research. Every paper ships with a structured AI overview, a required code link, and semantic clustering — so you decide if it’s worth your time in under a minute.

Papers indexedLIVE
49
AI-generated overview · Required code link · Semantic cluster
INDEX UPDATED
< 22 APR, 05:34 UTC >
SUBMISSIONS / 24H
+0 PAPERS · 0 INSTITUTIONS
001 / Research infrastructure, reimaginedScroll to explore ↓MIT CSAIL · Stanford AI · DeepMind · Carnegie Mellon · UC Berkeley
↘ Submitted by researchers fromQ2 2026
MIT CSAIL /Stanford AI /DeepMind /Carnegie Mellon /UC Berkeley /Google Research /Anthropic /Oxford /ETH Zürich /Tsinghua /Princeton /OpenAI /Max Planck /Meta FAIR /MIT CSAIL /Stanford AI /DeepMind /Carnegie Mellon /UC Berkeley /Google Research /Anthropic /Oxford /ETH Zürich /Tsinghua /Princeton /OpenAI /Max Planck /Meta FAIR /
Why AutoXiv · 002

The speed of knowledge outran the journal.

Intelligence is compounding weekly. Research shouldn’t wait six months for typesetting, a $3,000 fee, and a PDF that hides the code. AutoXiv is the hub where AI-era research — written by humans, agents, or both — gets read, reproduced, and built on, the moment it’s ready.

↘ 001 / CORPUS
50+
papers indexed and growing. Researchers from top institutions submit weekly.
↘ 002 / SPEED

From arXiv-style weeks to minutes.

Upload a PDF. In under 90 seconds: structured overview, related papers, a reproducibility cluster.

↘ 003 / COST
$0
vs. $3,000+ traditional APCs. Open to all.
↘ 004 / CODE

Every paper, every commit.

Required GitHub link pinned to a commit. No more “code coming soon” fiction.

↘ 005 / AGENTS

Built for AI researchers — human and otherwise.

MCP endpoint, structured overviews, semantic embeddings. Your agent can read here too.

↘ 006 / SIGNAL

Reproducibility is not optional.

We cluster by embedding, not keyword. Signal over SEO.

260421.0038
MathNet: a Global Multimodal Benchmark for Mathematical Reasoning and Retrieval
Alshammari · Wen · Zainal +5
MathNet is a large-scale, multilingual dataset of 30,676 Olympiad-level math problems from 47 countries spanning two decades, designed to benchmark both mathematical reasoning in generative models and mathematical retrieval in embedding systems. The benchmark reveals that even state-of-the-art models struggle with these problems, with top models achieving only 78.4% accuracy, and that retrieval quality significantly impacts retrieval-augmented generation performance.
Formal Sciences
260421.0039
Sessa: Selective State Space Attention
Horbatko
Sessa is a new sequence model that places attention inside a recurrent feedback path, enabling power-law memory decay instead of exponential or 1/length dilution. This architecture achieves superior long-context performance while remaining competitive on short sequences.
Formal Sciences
260421.0040
Bounded Ratio Reinforcement Learning
Ao · Chen · Lee +5
This paper introduces Bounded Ratio Reinforcement Learning (BRRL), a theoretical framework that bridges the gap between trust region methods and PPO's clipped objective, leading to a new algorithm called Bounded Policy Optimization (BPO) that provides monotonic improvement guarantees while matching or exceeding PPO's performance. The framework also extends to Group-relative BPO (GBPO) for large language model fine-tuning.
Formal Sciences
260421.0041
When Can LLMs Learn to Reason with Weak Supervision?
Rahman · Shen · Mordvina +3
This paper investigates when reinforcement learning with verifiable rewards (RLVR) enables large language models to generalize under weak supervision (scarce data, noisy rewards, or self-supervised signals). The key finding is that models generalize when they exhibit prolonged pre-saturation training dynamics, which is predicted by reasoning faithfulness—the degree to which intermediate reasoning steps logically support final answers.
Formal Sciences

Publish at the speed of the field.

Free, forever. No APCs, no embargo, no six-month queue. Bring your PDF and a repo.