NaturalReasoning: Reasoning in the Wild with 2.8M Challenging Questions
Abstract
Scaling reasoning capabilities beyond traditional domains such as math and coding is hindered by the lack of diverse and high-quality questions. To overcome this limitation, we introduce a scalable approach for generating diverse and challenging reasoning questions, accompanied by reference answers. We present NaturalReasoning, a comprehensive dataset comprising 2.8 million questions that span multiple domains, including STEM fields (e.g., Physics, Computer Science), Economics, Social Sciences, and more. We demonstrate the utility of the questions in NaturalReasoning through knowledge distillation experiments which show that NaturalReasoning can effectively elicit and transfer reasoning capabilities from a strong teacher model. Furthermore, we demonstrate that NaturalReasoning is also effective for unsupervised self-training using external reward models or self-rewarding.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- VersaPRM: Multi-Domain Process Reward Model via Synthetic Reasoning Data (2025)
- Expanding RL with Verifiable Rewards Across Diverse Domains (2025)
- Small Models Struggle to Learn from Strong Reasoners (2025)
- Fino1: On the Transferability of Reasoning Enhanced LLMs to Finance (2025)
- InfiR : Crafting Effective Small Language Models and Multimodal Small Language Models in Reasoning (2025)
- Enhancing LLM Reasoning with Iterative DPO: A Comprehensive Empirical Investigation (2025)
- DeepThink: Aligning Language Models with Domain-Specific User Intents (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper