view article Article Transformers backend integration in SGLang By marcsun13 and 4 others • Jun 23 • 51
view post Post 3125 Hunyuan-A13B 🔥 New MoE LLM by TencentHunyuan tencent/Hunyuan-A13B-Instruct✨80B total / 13B active params✨256K context window✨Dual-mode reasoning: fast & slow thinking✨Efficient inference (GQA + quantization) See translation 🔥 12 12 + Reply