SlimMoE: Structured Compression of Large MoE Models via Expert Slimming and Distillation Paper • 2506.18349 • Published Jun 23 • 13
COSMOS: A Hybrid Adaptive Optimizer for Memory-Efficient Training of LLMs Paper • 2502.17410 • Published Feb 24
LLMs Can Generate a Better Answer by Aggregating Their Own Responses Paper • 2503.04104 • Published Mar 6 • 1
SlimMoE: Structured Compression of Large MoE Models via Expert Slimming and Distillation Paper • 2506.18349 • Published Jun 23 • 13
KOROL: Learning Visualizable Object Feature with Koopman Operator Rollout for Manipulation Paper • 2407.00548 • Published Jun 29, 2024
Flow-of-Options: Diversified and Improved LLM Reasoning by Thinking Through Options Paper • 2502.12929 • Published Feb 18 • 7
Training Socially Aligned Language Models in Simulated Human Society Paper • 2305.16960 • Published May 26, 2023 • 3
Confidence Calibration and Rationalization for LLMs via Multi-Agent Deliberation Paper • 2404.09127 • Published Apr 14, 2024 • 2
Improved Generation of Synthetic Imaging Data Using Feature-Aligned Diffusion Paper • 2410.00731 • Published Oct 1, 2024
Creative Problem Solving in Large Language and Vision Models -- What Would it Take? Paper • 2405.01453 • Published May 2, 2024
HelpSteer2-Preference: Complementing Ratings with Preferences Paper • 2410.01257 • Published Oct 2, 2024 • 25
Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone Paper • 2404.14219 • Published Apr 22, 2024 • 257
GeorgiaTech/0.0005_llama_nodpo_3iters_bs128_531lr_oldtrl_iter_3 Text Generation • 8B • Updated May 13, 2024 • 2