Adding Evaluation Results
Browse filesThis is an automated PR created with https://huggingface.co/spaces/Weyaxi/open-llm-leaderboard-results-pr
The purpose of this PR is to add evaluation results from the Open LLM Leaderboard to your model card.
If you encounter any issues, please report them to https://huggingface.co/spaces/Weyaxi/open-llm-leaderboard-results-pr/discussions
    	
        README.md
    CHANGED
    
    | @@ -1,11 +1,114 @@ | |
| 1 | 
             
            ---
         | 
| 2 | 
             
            language:
         | 
| 3 | 
             
            - en
         | 
| 4 | 
            -
             | 
| 5 | 
             
            tags:
         | 
| 6 | 
             
            - meta
         | 
| 7 | 
             
            - llama-3
         | 
| 8 | 
            -
             | 
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
| 9 | 
             
            ---
         | 
| 10 | 
             
            <a href="https://www.gradient.ai" target="_blank"><img src="https://cdn-uploads.huggingface.co/production/uploads/655bb613e8a8971e89944f3e/TSa3V8YpoVagnTYgxiLaO.png" width="200"/></a>
         | 
| 11 |  | 
| @@ -673,3 +776,17 @@ Please see the Responsible Use Guide available at [http://llama.meta.com/respons | |
| 673 | 
             
            ## Contributors
         | 
| 674 |  | 
| 675 | 
             
            Aaditya Singh; Aaron Grattafiori; Abhimanyu Dubey; Abhinav Jauhri; Abhinav Pandey; Abhishek Kadian; Adam Kelsey; Adi Gangidi; Ahmad Al-Dahle; Ahuva Goldstand; Aiesha Letman; Ajay Menon; Akhil Mathur; Alan Schelten; Alex Vaughan; Amy Yang; Andrei Lupu; Andres Alvarado; Andrew Gallagher; Andrew Gu; Andrew Ho; Andrew Poulton; Andrew Ryan; Angela Fan; Ankit Ramchandani; Anthony Hartshorn; Archi Mitra; Archie Sravankumar; Artem Korenev; Arun Rao; Ashley Gabriel; Ashwin Bharambe; Assaf Eisenman; Aston Zhang; Aurelien Rodriguez; Austen Gregerson; Ava Spataru; Baptiste Roziere; Ben Maurer; Benjamin Leonhardi; Bernie Huang; Bhargavi Paranjape; Bing Liu; Binh Tang; Bobbie Chern; Brani Stojkovic; Brian Fuller; Catalina Mejia Arenas; Chao Zhou; Charlotte Caucheteux; Chaya Nayak; Ching-Hsiang Chu; Chloe Bi; Chris Cai; Chris Cox; Chris Marra; Chris McConnell; Christian Keller; Christoph Feichtenhofer; Christophe Touret; Chunyang Wu; Corinne Wong; Cristian Canton Ferrer; Damien Allonsius; Daniel Kreymer; Daniel Haziza; Daniel Li; Danielle Pintz; Danny Livshits; Danny Wyatt; David Adkins; David Esiobu; David Xu; Davide Testuggine; Delia David; Devi Parikh; Dhruv Choudhary; Dhruv Mahajan; Diana Liskovich; Diego Garcia-Olano; Diego Perino; Dieuwke Hupkes; Dingkang Wang; Dustin Holland; Egor Lakomkin; Elina Lobanova; Xiaoqing Ellen Tan; Emily Dinan; Eric Smith; Erik Brinkman; Esteban Arcaute; Filip Radenovic; Firat Ozgenel; Francesco Caggioni; Frank Seide; Frank Zhang; Gabriel Synnaeve; Gabriella Schwarz; Gabrielle Lee; Gada Badeer; Georgia Anderson; Graeme Nail; Gregoire Mialon; Guan Pang; Guillem Cucurell; Hailey Nguyen; Hannah Korevaar; Hannah Wang; Haroun Habeeb; Harrison Rudolph; Henry Aspegren; Hu Xu; Hugo Touvron; Iga Kozlowska; Igor Molybog; Igor Tufanov; Iliyan Zarov; Imanol Arrieta Ibarra; Irina-Elena Veliche; Isabel Kloumann; Ishan Misra; Ivan Evtimov; Jacob Xu; Jade Copet; Jake Weissman; Jan Geffert; Jana Vranes; Japhet Asher; Jason Park; Jay Mahadeokar; Jean-Baptiste Gaya; Jeet Shah; Jelmer van der Linde; Jennifer Chan; Jenny Hong; Jenya Lee; Jeremy Fu; Jeremy Teboul; Jianfeng Chi; Jianyu Huang; Jie Wang; Jiecao Yu; Joanna Bitton; Joe Spisak; Joelle Pineau; Jon Carvill; Jongsoo Park; Joseph Rocca; Joshua Johnstun; Junteng Jia; Kalyan Vasuden Alwala; Kam Hou U; Kate Plawiak; Kartikeya Upasani; Kaushik Veeraraghavan; Ke Li; Kenneth Heafield; Kevin Stone; Khalid El-Arini; Krithika Iyer; Kshitiz Malik; Kuenley Chiu; Kunal Bhalla; Kyle Huang; Lakshya Garg; Lauren Rantala-Yeary; Laurens van der Maaten; Lawrence Chen; Leandro Silva; Lee Bell; Lei Zhang; Liang Tan; Louis Martin; Lovish Madaan; Luca Wehrstedt; Lukas Blecher; Luke de Oliveira; Madeline Muzzi; Madian Khabsa; Manav Avlani; Mannat Singh; Manohar Paluri; Mark Zuckerberg; Marcin Kardas; Martynas Mankus; Mathew Oldham; Mathieu Rita; Matthew Lennie; Maya Pavlova; Meghan Keneally; Melanie Kambadur; Mihir Patel; Mikayel Samvelyan; Mike Clark; Mike Lewis; Min Si; Mitesh Kumar Singh; Mo Metanat; Mona Hassan; Naman Goyal; Narjes Torabi; Nicolas Usunier; Nikolay Bashlykov; Nikolay Bogoychev; Niladri Chatterji; Ning Dong; Oliver Aobo Yang; Olivier Duchenne; Onur Celebi; Parth Parekh; Patrick Alrassy; Paul Saab; Pavan Balaji; Pedro Rittner; Pengchuan Zhang; Pengwei Li; Petar Vasic; Peter Weng; Polina Zvyagina; Prajjwal Bhargava; Pratik Dubal; Praveen Krishnan; Punit Singh Koura; Qing He; Rachel Rodriguez; Ragavan Srinivasan; Rahul Mitra; Ramon Calderer; Raymond Li; Robert Stojnic; Roberta Raileanu; Robin Battey; Rocky Wang; Rohit Girdhar; Rohit Patel; Romain Sauvestre; Ronnie Polidoro; Roshan Sumbaly; Ross Taylor; Ruan Silva; Rui Hou; Rui Wang; Russ Howes; Ruty Rinott; Saghar Hosseini; Sai Jayesh Bondu; Samyak Datta; Sanjay Singh; Sara Chugh; Sargun Dhillon; Satadru Pan; Sean Bell; Sergey Edunov; Shaoliang Nie; Sharan Narang; Sharath Raparthy; Shaun Lindsay; Sheng Feng; Sheng Shen; Shenghao Lin; Shiva Shankar; Shruti Bhosale; Shun Zhang; Simon Vandenhende; Sinong Wang; Seohyun Sonia Kim; Soumya Batra; Sten Sootla; Steve Kehoe; Suchin Gururangan; Sumit Gupta; Sunny Virk; Sydney Borodinsky; Tamar Glaser; Tamar Herman; Tamara Best; Tara Fowler; Thomas Georgiou; Thomas Scialom; Tianhe Li; Todor Mihaylov; Tong Xiao; Ujjwal Karn; Vedanuj Goswami; Vibhor Gupta; Vignesh Ramanathan; Viktor Kerkez; Vinay Satish Kumar; Vincent Gonguet; Vish Vogeti; Vlad Poenaru; Vlad Tiberiu Mihailescu; Vladan Petrovic; Vladimir Ivanov; Wei Li; Weiwei Chu; Wenhan Xiong; Wenyin Fu; Wes Bouaziz; Whitney Meers; Will Constable; Xavier Martinet; Xiaojian Wu; Xinbo Gao; Xinfeng Xie; Xuchao Jia; Yaelle Goldschlag; Yann LeCun; Yashesh Gaur; Yasmine Babaei; Ye Qi; Yenda Li; Yi Wen; Yiwen Song; Youngjin Nam; Yuchen Hao; Yuchen Zhang; Yun Wang; Yuning Mao; Yuzi He; Zacharie Delpierre Coudert; Zachary DeVito; Zahra Hankir; Zhaoduo Wen; Zheng Yan; Zhengxing Chen; Zhenyu Yang; Zoe Papakipos
         | 
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
             
            ---
         | 
| 2 | 
             
            language:
         | 
| 3 | 
             
            - en
         | 
| 4 | 
            +
            license: llama3
         | 
| 5 | 
             
            tags:
         | 
| 6 | 
             
            - meta
         | 
| 7 | 
             
            - llama-3
         | 
| 8 | 
            +
            pipeline_tag: text-generation
         | 
| 9 | 
            +
            model-index:
         | 
| 10 | 
            +
            - name: Llama-3-70B-Instruct-Gradient-262k
         | 
| 11 | 
            +
              results:
         | 
| 12 | 
            +
              - task:
         | 
| 13 | 
            +
                  type: text-generation
         | 
| 14 | 
            +
                  name: Text Generation
         | 
| 15 | 
            +
                dataset:
         | 
| 16 | 
            +
                  name: AI2 Reasoning Challenge (25-Shot)
         | 
| 17 | 
            +
                  type: ai2_arc
         | 
| 18 | 
            +
                  config: ARC-Challenge
         | 
| 19 | 
            +
                  split: test
         | 
| 20 | 
            +
                  args:
         | 
| 21 | 
            +
                    num_few_shot: 25
         | 
| 22 | 
            +
                metrics:
         | 
| 23 | 
            +
                - type: acc_norm
         | 
| 24 | 
            +
                  value: 67.58
         | 
| 25 | 
            +
                  name: normalized accuracy
         | 
| 26 | 
            +
                source:
         | 
| 27 | 
            +
                  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=gradientai/Llama-3-70B-Instruct-Gradient-262k
         | 
| 28 | 
            +
                  name: Open LLM Leaderboard
         | 
| 29 | 
            +
              - task:
         | 
| 30 | 
            +
                  type: text-generation
         | 
| 31 | 
            +
                  name: Text Generation
         | 
| 32 | 
            +
                dataset:
         | 
| 33 | 
            +
                  name: HellaSwag (10-Shot)
         | 
| 34 | 
            +
                  type: hellaswag
         | 
| 35 | 
            +
                  split: validation
         | 
| 36 | 
            +
                  args:
         | 
| 37 | 
            +
                    num_few_shot: 10
         | 
| 38 | 
            +
                metrics:
         | 
| 39 | 
            +
                - type: acc_norm
         | 
| 40 | 
            +
                  value: 86.4
         | 
| 41 | 
            +
                  name: normalized accuracy
         | 
| 42 | 
            +
                source:
         | 
| 43 | 
            +
                  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=gradientai/Llama-3-70B-Instruct-Gradient-262k
         | 
| 44 | 
            +
                  name: Open LLM Leaderboard
         | 
| 45 | 
            +
              - task:
         | 
| 46 | 
            +
                  type: text-generation
         | 
| 47 | 
            +
                  name: Text Generation
         | 
| 48 | 
            +
                dataset:
         | 
| 49 | 
            +
                  name: MMLU (5-Shot)
         | 
| 50 | 
            +
                  type: cais/mmlu
         | 
| 51 | 
            +
                  config: all
         | 
| 52 | 
            +
                  split: test
         | 
| 53 | 
            +
                  args:
         | 
| 54 | 
            +
                    num_few_shot: 5
         | 
| 55 | 
            +
                metrics:
         | 
| 56 | 
            +
                - type: acc
         | 
| 57 | 
            +
                  value: 77.19
         | 
| 58 | 
            +
                  name: accuracy
         | 
| 59 | 
            +
                source:
         | 
| 60 | 
            +
                  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=gradientai/Llama-3-70B-Instruct-Gradient-262k
         | 
| 61 | 
            +
                  name: Open LLM Leaderboard
         | 
| 62 | 
            +
              - task:
         | 
| 63 | 
            +
                  type: text-generation
         | 
| 64 | 
            +
                  name: Text Generation
         | 
| 65 | 
            +
                dataset:
         | 
| 66 | 
            +
                  name: TruthfulQA (0-shot)
         | 
| 67 | 
            +
                  type: truthful_qa
         | 
| 68 | 
            +
                  config: multiple_choice
         | 
| 69 | 
            +
                  split: validation
         | 
| 70 | 
            +
                  args:
         | 
| 71 | 
            +
                    num_few_shot: 0
         | 
| 72 | 
            +
                metrics:
         | 
| 73 | 
            +
                - type: mc2
         | 
| 74 | 
            +
                  value: 54.68
         | 
| 75 | 
            +
                source:
         | 
| 76 | 
            +
                  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=gradientai/Llama-3-70B-Instruct-Gradient-262k
         | 
| 77 | 
            +
                  name: Open LLM Leaderboard
         | 
| 78 | 
            +
              - task:
         | 
| 79 | 
            +
                  type: text-generation
         | 
| 80 | 
            +
                  name: Text Generation
         | 
| 81 | 
            +
                dataset:
         | 
| 82 | 
            +
                  name: Winogrande (5-shot)
         | 
| 83 | 
            +
                  type: winogrande
         | 
| 84 | 
            +
                  config: winogrande_xl
         | 
| 85 | 
            +
                  split: validation
         | 
| 86 | 
            +
                  args:
         | 
| 87 | 
            +
                    num_few_shot: 5
         | 
| 88 | 
            +
                metrics:
         | 
| 89 | 
            +
                - type: acc
         | 
| 90 | 
            +
                  value: 83.98
         | 
| 91 | 
            +
                  name: accuracy
         | 
| 92 | 
            +
                source:
         | 
| 93 | 
            +
                  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=gradientai/Llama-3-70B-Instruct-Gradient-262k
         | 
| 94 | 
            +
                  name: Open LLM Leaderboard
         | 
| 95 | 
            +
              - task:
         | 
| 96 | 
            +
                  type: text-generation
         | 
| 97 | 
            +
                  name: Text Generation
         | 
| 98 | 
            +
                dataset:
         | 
| 99 | 
            +
                  name: GSM8k (5-shot)
         | 
| 100 | 
            +
                  type: gsm8k
         | 
| 101 | 
            +
                  config: main
         | 
| 102 | 
            +
                  split: test
         | 
| 103 | 
            +
                  args:
         | 
| 104 | 
            +
                    num_few_shot: 5
         | 
| 105 | 
            +
                metrics:
         | 
| 106 | 
            +
                - type: acc
         | 
| 107 | 
            +
                  value: 80.44
         | 
| 108 | 
            +
                  name: accuracy
         | 
| 109 | 
            +
                source:
         | 
| 110 | 
            +
                  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=gradientai/Llama-3-70B-Instruct-Gradient-262k
         | 
| 111 | 
            +
                  name: Open LLM Leaderboard
         | 
| 112 | 
             
            ---
         | 
| 113 | 
             
            <a href="https://www.gradient.ai" target="_blank"><img src="https://cdn-uploads.huggingface.co/production/uploads/655bb613e8a8971e89944f3e/TSa3V8YpoVagnTYgxiLaO.png" width="200"/></a>
         | 
| 114 |  | 
|  | |
| 776 | 
             
            ## Contributors
         | 
| 777 |  | 
| 778 | 
             
            Aaditya Singh; Aaron Grattafiori; Abhimanyu Dubey; Abhinav Jauhri; Abhinav Pandey; Abhishek Kadian; Adam Kelsey; Adi Gangidi; Ahmad Al-Dahle; Ahuva Goldstand; Aiesha Letman; Ajay Menon; Akhil Mathur; Alan Schelten; Alex Vaughan; Amy Yang; Andrei Lupu; Andres Alvarado; Andrew Gallagher; Andrew Gu; Andrew Ho; Andrew Poulton; Andrew Ryan; Angela Fan; Ankit Ramchandani; Anthony Hartshorn; Archi Mitra; Archie Sravankumar; Artem Korenev; Arun Rao; Ashley Gabriel; Ashwin Bharambe; Assaf Eisenman; Aston Zhang; Aurelien Rodriguez; Austen Gregerson; Ava Spataru; Baptiste Roziere; Ben Maurer; Benjamin Leonhardi; Bernie Huang; Bhargavi Paranjape; Bing Liu; Binh Tang; Bobbie Chern; Brani Stojkovic; Brian Fuller; Catalina Mejia Arenas; Chao Zhou; Charlotte Caucheteux; Chaya Nayak; Ching-Hsiang Chu; Chloe Bi; Chris Cai; Chris Cox; Chris Marra; Chris McConnell; Christian Keller; Christoph Feichtenhofer; Christophe Touret; Chunyang Wu; Corinne Wong; Cristian Canton Ferrer; Damien Allonsius; Daniel Kreymer; Daniel Haziza; Daniel Li; Danielle Pintz; Danny Livshits; Danny Wyatt; David Adkins; David Esiobu; David Xu; Davide Testuggine; Delia David; Devi Parikh; Dhruv Choudhary; Dhruv Mahajan; Diana Liskovich; Diego Garcia-Olano; Diego Perino; Dieuwke Hupkes; Dingkang Wang; Dustin Holland; Egor Lakomkin; Elina Lobanova; Xiaoqing Ellen Tan; Emily Dinan; Eric Smith; Erik Brinkman; Esteban Arcaute; Filip Radenovic; Firat Ozgenel; Francesco Caggioni; Frank Seide; Frank Zhang; Gabriel Synnaeve; Gabriella Schwarz; Gabrielle Lee; Gada Badeer; Georgia Anderson; Graeme Nail; Gregoire Mialon; Guan Pang; Guillem Cucurell; Hailey Nguyen; Hannah Korevaar; Hannah Wang; Haroun Habeeb; Harrison Rudolph; Henry Aspegren; Hu Xu; Hugo Touvron; Iga Kozlowska; Igor Molybog; Igor Tufanov; Iliyan Zarov; Imanol Arrieta Ibarra; Irina-Elena Veliche; Isabel Kloumann; Ishan Misra; Ivan Evtimov; Jacob Xu; Jade Copet; Jake Weissman; Jan Geffert; Jana Vranes; Japhet Asher; Jason Park; Jay Mahadeokar; Jean-Baptiste Gaya; Jeet Shah; Jelmer van der Linde; Jennifer Chan; Jenny Hong; Jenya Lee; Jeremy Fu; Jeremy Teboul; Jianfeng Chi; Jianyu Huang; Jie Wang; Jiecao Yu; Joanna Bitton; Joe Spisak; Joelle Pineau; Jon Carvill; Jongsoo Park; Joseph Rocca; Joshua Johnstun; Junteng Jia; Kalyan Vasuden Alwala; Kam Hou U; Kate Plawiak; Kartikeya Upasani; Kaushik Veeraraghavan; Ke Li; Kenneth Heafield; Kevin Stone; Khalid El-Arini; Krithika Iyer; Kshitiz Malik; Kuenley Chiu; Kunal Bhalla; Kyle Huang; Lakshya Garg; Lauren Rantala-Yeary; Laurens van der Maaten; Lawrence Chen; Leandro Silva; Lee Bell; Lei Zhang; Liang Tan; Louis Martin; Lovish Madaan; Luca Wehrstedt; Lukas Blecher; Luke de Oliveira; Madeline Muzzi; Madian Khabsa; Manav Avlani; Mannat Singh; Manohar Paluri; Mark Zuckerberg; Marcin Kardas; Martynas Mankus; Mathew Oldham; Mathieu Rita; Matthew Lennie; Maya Pavlova; Meghan Keneally; Melanie Kambadur; Mihir Patel; Mikayel Samvelyan; Mike Clark; Mike Lewis; Min Si; Mitesh Kumar Singh; Mo Metanat; Mona Hassan; Naman Goyal; Narjes Torabi; Nicolas Usunier; Nikolay Bashlykov; Nikolay Bogoychev; Niladri Chatterji; Ning Dong; Oliver Aobo Yang; Olivier Duchenne; Onur Celebi; Parth Parekh; Patrick Alrassy; Paul Saab; Pavan Balaji; Pedro Rittner; Pengchuan Zhang; Pengwei Li; Petar Vasic; Peter Weng; Polina Zvyagina; Prajjwal Bhargava; Pratik Dubal; Praveen Krishnan; Punit Singh Koura; Qing He; Rachel Rodriguez; Ragavan Srinivasan; Rahul Mitra; Ramon Calderer; Raymond Li; Robert Stojnic; Roberta Raileanu; Robin Battey; Rocky Wang; Rohit Girdhar; Rohit Patel; Romain Sauvestre; Ronnie Polidoro; Roshan Sumbaly; Ross Taylor; Ruan Silva; Rui Hou; Rui Wang; Russ Howes; Ruty Rinott; Saghar Hosseini; Sai Jayesh Bondu; Samyak Datta; Sanjay Singh; Sara Chugh; Sargun Dhillon; Satadru Pan; Sean Bell; Sergey Edunov; Shaoliang Nie; Sharan Narang; Sharath Raparthy; Shaun Lindsay; Sheng Feng; Sheng Shen; Shenghao Lin; Shiva Shankar; Shruti Bhosale; Shun Zhang; Simon Vandenhende; Sinong Wang; Seohyun Sonia Kim; Soumya Batra; Sten Sootla; Steve Kehoe; Suchin Gururangan; Sumit Gupta; Sunny Virk; Sydney Borodinsky; Tamar Glaser; Tamar Herman; Tamara Best; Tara Fowler; Thomas Georgiou; Thomas Scialom; Tianhe Li; Todor Mihaylov; Tong Xiao; Ujjwal Karn; Vedanuj Goswami; Vibhor Gupta; Vignesh Ramanathan; Viktor Kerkez; Vinay Satish Kumar; Vincent Gonguet; Vish Vogeti; Vlad Poenaru; Vlad Tiberiu Mihailescu; Vladan Petrovic; Vladimir Ivanov; Wei Li; Weiwei Chu; Wenhan Xiong; Wenyin Fu; Wes Bouaziz; Whitney Meers; Will Constable; Xavier Martinet; Xiaojian Wu; Xinbo Gao; Xinfeng Xie; Xuchao Jia; Yaelle Goldschlag; Yann LeCun; Yashesh Gaur; Yasmine Babaei; Ye Qi; Yenda Li; Yi Wen; Yiwen Song; Youngjin Nam; Yuchen Hao; Yuchen Zhang; Yun Wang; Yuning Mao; Yuzi He; Zacharie Delpierre Coudert; Zachary DeVito; Zahra Hankir; Zhaoduo Wen; Zheng Yan; Zhengxing Chen; Zhenyu Yang; Zoe Papakipos
         | 
| 779 | 
            +
             | 
| 780 | 
            +
            # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
         | 
| 781 | 
            +
            Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_gradientai__Llama-3-70B-Instruct-Gradient-262k)
         | 
| 782 | 
            +
             | 
| 783 | 
            +
            |             Metric              |Value|
         | 
| 784 | 
            +
            |---------------------------------|----:|
         | 
| 785 | 
            +
            |Avg.                             |75.04|
         | 
| 786 | 
            +
            |AI2 Reasoning Challenge (25-Shot)|67.58|
         | 
| 787 | 
            +
            |HellaSwag (10-Shot)              |86.40|
         | 
| 788 | 
            +
            |MMLU (5-Shot)                    |77.19|
         | 
| 789 | 
            +
            |TruthfulQA (0-shot)              |54.68|
         | 
| 790 | 
            +
            |Winogrande (5-shot)              |83.98|
         | 
| 791 | 
            +
            |GSM8k (5-shot)                   |80.44|
         | 
| 792 | 
            +
             | 

