top of page
WebThinker: Autonomous Web Exploration and Report Writing for LRMs
- Methodology: WebThinker introduces a framework equipping LRMs with a Deep Web Explorer for autonomous web search/navigation and an Autonomous Think-Search-and-Draft strategy for interleaving reasoning, information gathering, and report writing. It employs online Direct Preference Optimization (DPO) for RL-based training to improve tool usage.
- Performance: WebThinker-32B-RL achieves state-of-the-art results among 32B models on complex reasoning benchmarks (GPQA: 70.7%, HLE: 15.8%), outperforming retrieval-augmented and proprietary systems, and excels in scientific report writing on the Glaive dataset (scoring 8.1 in average quality metrics).
- RL Refinement: RL-trained versions consistently outperform base counterparts, demonstrating that iterative preference-based learning significantly enhances reasoning-tool coordination.
- Ablation Studies: Removing the Deep Web Explorer or automatic report drafting significantly degrades performance, validating the necessity of these components.
- Implementation Details: The system uses QwQ-32B or DeepSeek-R1-Distilled-Qwen models as the backbone, Qwen2.5-Instruct as assistant models, Bing Web Search API for search, and Crawl4AI for web page content. Training involves 2 iterations of online DPO.
- _According to additional sources:_ The preference data construction for DPO considers overall correctness/quality, tool efficiency (fewer tool calls), and thinking conciseness (output length ratio).
Source:
bottom of page