Skip to content
Go back 2505.04588 arXiv logo

ZeroSearch: Incentivize the Search Capability of LLMs without Searching

Published:  at  06:16 PM
86.09 🤔

ZEROSEARCH introduces a reinforcement learning framework that enhances LLMs’ search capabilities by simulating search engines with fine-tuned LLMs, achieving performance comparable to or better than real search engines without API costs through a curriculum-based rollout strategy.

Reinforcement Learning, Large Language Model, Retrieval-Augmented Generation, Curriculum Learning, Simulation, Reasoning

Hao Sun, Zile Qiao, Jiayan Guo, Xuanbo Fan, Yingyan Hou, Yong Jiang, Pengjun Xie, Fei Huang, Yan Zhang

Tongyi Lab, Alibaba Group

Generated by grok-3

Background Problem

Large Language Models (LLMs) excel in various tasks but are limited by static knowledge from pretraining, often leading to hallucinations or outdated information. Enhancing their ability to search for external information is crucial for accurate responses. Existing methods like Retrieval-Augmented Generation (RAG) and reinforcement learning (RL) with real search engines face challenges such as unpredictable document quality from live searches and high API costs due to frequent rollouts. ZEROSEARCH addresses these issues by proposing a novel RL framework that trains LLMs to search without real search engine interaction, aiming to reduce costs and control document quality while maintaining or improving performance.

Method

ZEROSEARCH is a reinforcement learning framework that uses an LLM to simulate a search engine, eliminating the need for real-world API calls. The core idea is to transform an LLM into a retrieval module through lightweight supervised fine-tuning (SFT), enabling it to generate both relevant and noisy documents based on query prompts. The process involves: (1) Collecting interaction trajectories with a real search engine to label positive (useful) and negative (noisy) documents, then fine-tuning the simulation LLM to mimic these outputs by adjusting prompt wording. (2) During RL training, a curriculum-based rollout strategy is applied where the probability of generating noisy documents increases over time using a defined function pi=ps+bi/m1b1(peps)p_i = p_s + \frac{b^{i/m} - 1}{b - 1} (p_e - p_s), allowing the policy model to adapt from easy to challenging retrieval scenarios. (3) The policy model interacts in a structured template with reasoning, search, and answer stages, optimized using RL algorithms like PPO or GRPO, with a reward function based on F1 score to prevent reward hacking. Loss masking ensures gradients are computed only for the policy model’s outputs, stabilizing training.

Experiment

Experiments were conducted on diverse question-answering benchmarks, including single-hop (NQ, TriviaQA, PopQA) and multi-hop (HotpotQA, 2WikiMultiHopQA, Musique, Bamboogle) datasets, using Exact Match (EM) as the evaluation metric. Models tested include Qwen-2.5 (3B, 7B, 14B) and LLaMA-3.2 (3B) in both base and instruction-tuned variants. The setup merged NQ and HotpotQA for training, evaluating in-domain and out-of-domain performance against baselines like vanilla prompting, advanced RAG, and RL methods (e.g., Search-R1 with real search engines). Simulation LLMs were fine-tuned at different sizes (3B to 14B), and RL algorithms (PPO, GRPO) were compared. Results show ZEROSEARCH consistently outperforms baselines, with a 7B simulation model matching Google Search performance and a 14B model surpassing it (e.g., average EM of 33.97 vs. 32.47 for Google on Qwen-2.5-3B). The curriculum strategy proved effective, as easy-to-hard progression outperformed reverse curriculum. However, while results are promising, the reliance on EM might not fully reflect real-world search nuances, and the computational cost of larger simulation models (e.g., 14B) wasn’t thoroughly quantified against performance gains. The setup is comprehensive for QA tasks but lacks diversity in non-QA search scenarios, potentially limiting generalizability.

Further Thoughts

The ZEROSEARCH framework opens intriguing avenues for reducing dependency on external APIs in LLM training, a significant step toward cost-effective scaling of AI research. However, I’m concerned about the long-term implications of training in a simulated environment—does it risk creating models overly tuned to artificial data distributions, potentially failing in real-world, unpredictable search scenarios? This could be explored by testing ZEROSEARCH-trained models against live search engines in dynamic, unseen domains. Additionally, the curriculum learning approach, while effective in controlled settings, might benefit from adaptive mechanisms that adjust noise levels based on model performance rather than a predefined schedule, potentially improving robustness. I also see a connection to privacy-preserving machine learning; simulating search could minimize data exposure risks inherent in real API interactions, an angle worth investigating further. Finally, comparing ZEROSEARCH’s simulation strategy with synthetic data generation techniques in other domains (e.g., image synthesis for vision models) could reveal cross-disciplinary insights into balancing realism and cost in AI training pipelines.



Previous Post
Dynamic Fisher-weighted Model Merging via Bayesian Optimization
Next Post
Investigating Task Arithmetic for Zero-Shot Information Retrieval