David Quispe's picture

David Quispe PRO

daqc

AI & ML interests

Education

Recent Activity

Organizations

Hackathon Somos NLP 2023: Los LLMs hablan EspaΓ±ol's profile picture SomosNLP's profile picture Blog-explorers's profile picture MLX Community's profile picture PlawLabs's profile picture SyndroBytes's profile picture Hackathon SomosNLP 2025's profile picture AI Starter Pack's profile picture Quechua LLMs on Hugging Face's profile picture

daqc's activity

upvoted 2 articles 7 days ago
view article
Article

NVIDIA's GTC 2025 Announcement for Physical AI Developers: New Open Models and Datasets

β€’ 35
view article
Article

Hugging Face to sell open-source robots thanks to Pollen Robotics acquisition πŸ€–

β€’ 38
upvoted 2 articles 17 days ago
view article
Article

Welcome Gemma 3: Google's all new multimodal, multilingual, long context open LLM

β€’ 398
view article
Article

Welcome Llama 4 Maverick & Scout on Hugging Face!

β€’ 140
reacted to lewtun's post with πŸ”₯ 27 days ago
view post
Post
2421
Introducing OlympicCoder: a series of open reasoning models that can solve olympiad-level programming problems πŸ§‘β€πŸ’»

- 7B open-r1/OlympicCoder-7B
- 32B open-r1/OlympicCoder-32B

We find that OlympicCoder models outperform Claude 3.7 Sonnet, as well as others over 100x larger πŸ’ͺ

Together with the models, we are releasing:

πŸ“ŠCodeForces-CoTs: new dataset of code problems from the most popular competitive coding platform, with R1 traces in C++ and Python open-r1/codeforces-cots

πŸ† IOI'2024: a new benchmark of VERY hard programming problems where even frontier models struggle to match human performance open-r1/ioi

For links to the models and datasets, check out our latest progress report from Open R1: https://huggingface.co/blog/open-r1/update-3
  • 1 reply
Β·
reacted to lewtun's post with ❀️ 27 days ago
view post
Post
5139
Introducing OpenR1-Math-220k!

open-r1/OpenR1-Math-220k

The community has been busy distilling DeepSeek-R1 from inference providers, but we decided to have a go at doing it ourselves from scratch πŸ’ͺ

What’s new compared to existing reasoning datasets?

β™Ύ Based on AI-MO/NuminaMath-1.5: we focus on math reasoning traces and generate answers for problems in NuminaMath 1.5, an improved version of the popular NuminaMath-CoT dataset.

🐳 800k R1 reasoning traces: We generate two answers for 400k problems using DeepSeek R1. The filtered dataset contains 220k problems with correct reasoning traces.

πŸ“€ 512 H100s running locally: Instead of relying on an API, we leverage vLLM and SGLang to run generations locally on our science cluster, generating 180k reasoning traces per day.

⏳ Automated filtering: We apply Math Verify to only retain problems with at least one correct answer. We also leverage Llama3.3-70B-Instruct as a judge to retrieve more correct examples (e.g for cases with malformed answers that can’t be verified with a rules-based parser)

πŸ“Š We match the performance of DeepSeek-Distill-Qwen-7B by finetuning Qwen-7B-Math-Instruct on our dataset.

πŸ”Ž Read our blog post for all the nitty gritty details: https://huggingface.co/blog/open-r1/update-2