view post Post 2342 I've run the open llm leaderboard evaluations + hellaswag on deepseek-ai/DeepSeek-R1-Distill-Llama-8B and compared to meta-llama/Llama-3.1-8B-Instruct and at first glance R1 do not beat Llama overall.If anyone wants to double check the results are posted here: https://github.com/csabakecskemeti/lm_eval_resultsAm I made some mistake, or (at least this distilled version) not as good/better than the competition?I'll run the same on the Qwen 7B distilled version too. See translation 7 replies · 👀 6 6 + Reply
Mutual Reasoning Makes Smaller LLMs Stronger Problem-Solvers Paper • 2408.06195 • Published Aug 12, 2024 • 73