

Learning Adaptive Parallel Reasoning with Language Models
AI & ML interests
None defined yet.
Recent Activity
Learning Adaptive Parallel Reasoning with Language Models
Jiayi Pan*,
Xiuyu Li*,
Long Lian*,
Charlie Victor Snell,
Yifei Zhou,
Adam Yala,
Trevor Darrell,
Kurt Keutzer,
Alane Suhr
UC Berkeley and UCSF * Equal Contribution
📃 Paper • 💻 Code • 🤗 Data & Models
TL;DR:
We present Adaptive Parallel Reasoning (APR), a novel framework that enables language models to learn to orchestrate both serialized and parallel computations. APR trains language models to use spawn()
and join()
operations through end-to-end supervised training and reinforcement learning, allowing models to dynamically orchestrate their own computational workflows.
APR efficiently distributes compute, reduces latency, overcomes context window limits, and achieves state‑of‑the‑art performance on complex reasoning tasks (e.g., 83.4% vs. 60.0% accuracy at 4K context on Countdown).
models
6


Parallel-Reasoning/llama-apr-beam10_subbeam15_num_subcall_cond

Parallel-Reasoning/llama-apr-beam10_subbeam15

Parallel-Reasoning/llama-apr_grpo

Parallel-Reasoning/llama-apr_cond10_grpo
