We are excited to contribute to the best 32B reasoning model- The post-training data was created with our Bespoke Curator tool.
Announcing OpenThinker-32B 🧠 The best open-data reasoning model distilled from DeepSeek-R1! Our results show that large, carefully curated datasets with verified R1 annotations produce SoTA reasoning models. OpenThinker-32B outperforms **all** 32B reasoning models including DeepSeek-R1-Distill-Qwen-32B (a closed data model) in MATH500 and GPQA Diamond, and shows similar performance in other benchmarks. Blog Post: https://lnkd.in/eD7gkK7b Model: https://lnkd.in/edGxGSjY Dataset: https://lnkd.in/eNpTsmsQ Data Curation Code: https://lnkd.in/eBfGvQSq Evaluation Code: https://lnkd.in/eErj2vbp An incredible team effort from Negin Raoof, Etash Guha, Trung Vu, Sedrick Keh, Marianna Nezhurina, Zaid Khan, George Smyrnis, Shreyas Pimpalgaonkar, Hritik Bansal, Jean Mercat, Mike Merrill, Niklas Muennighoff, Jenia Jitsev, Mahesh (Maheswaran) Sathiamoorthy, Alex Dimakis, Yejin Choi, Tatsu Hashimoto, and Ludwig Schmidt