The hype is real: a mysterious gpt2-chatbot model has appeared on the LLM Arena Leaderboard ๐. It seems to be at least on par with the top performing models (closed and open).
To try it out: https://chat.lmsys.org/ -> then click on the Direct Chat tab and select gpt2-chatbot.
How Robust Is Your Model in Complex Code Generation Tasks? ๐ค
We've launched the PECC benchmark to challenge chat models in code generation, drawing from the Advent of Code for programming tasks and the Euler Project for math-heavy challenges. This new task tests models with problems presented in both detailed prose and concise "leet code" styles, evaluating their ability to understand and solve complex coding issues and math problem in chat-based interactions.
It seems that the Claude 3 models outperforme ChatGPT: Model / Avg. (pass@3) Claude 3 Haiku / 27.67 GPT-3.5-Turbo / 23.75 Mixtral-8x22B-Instruct-v0.1 / 8.35
Some tokens are more relevant than others, and some are mostly noise (just look up the history of ๐๐ฐ๐ญ๐ช๐ฅ๐๐ฐ๐ญ๐ฅ๐๐ข๐จ๐ช๐ฌ๐ข๐ณ๐ฑ).
So this paper introduces ๐ฆ๐ฒ๐น๐ฒ๐ฐ๐๐ถ๐๐ฒ ๐๐ฎ๐ป๐ด๐๐ฎ๐ด๐ฒ ๐ ๐ผ๐ฑ๐ฒ๐น๐ถ๐ป๐ด, which is actually really simple: โก๏ธ A specific metric measures the relevance of each token. Then during training, only the top k% tokens for this relevance metric count in the loss calculation.
Authors test this method by training models on the difficult MATH dataset (only competition mathematics problems).
โก๏ธ Their technique seems like a new must-do in LLM training: Training is much faster and reaches an impressive performance!
๐๐๐ฌ๐ฎ๐ฅ๐ญ๐ฌ: โ โฑ๏ธ Training is x5 to x10 faster to reach equivalent performance compared to standard language modeling. โ ๐ช Their 1B model achieves close to GPT4 Chain-of-Thought performance on MATH! โ ๐ Their 7B model match performance of the state-of-the-art DeepSeek for the same size, while trained on only 3% of tokens
๐๐๐๐ข๐ญ๐ข๐จ๐ง๐๐ฅ ๐ข๐ง๐ฌ๐ข๐ ๐ก๐ญ๐ฌ ๐ก โ Datasets used for pre-training, even after pre-filtering, still contain a large proportion of noisy tokens ๐ โ Authors show that when you reduce loss on noisy tokens, you actually reduce accuracy (Figure 7). So Selective Language Modeling seems fundamental! โ