Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up
Testerpce 's Collections
LLM judge
Test time
3D
Physics and operators
Materials and structures
Vision Language Action models
Vision
World model
Code
Compression
Data
Process Reward Modelling
Memory
SAE
Applications and Uses
Theory and Representation learning
Adversarial
Graph
Multimodal
Search
Interpretable
Diversity
Diffusion
Self correction
Information_retrieval
Speech
Attention
Synthetic data
Agent
MoE
RAG
Markov chain
Prompt papers
Planning
Sparsity
Multilingual
State space LLM
Partial layer training LLMs
Reasoning
Evaluation
Fine tuning
Math
Dataset and Data processing
Style transfer
Video understanding
Reinforcement learning
Long context
Knowledge

MoE

updated 28 days ago
Upvote
-

  • Your Mixture-of-Experts LLM Is Secretly an Embedding Model For Free

    Paper • 2410.10814 • Published Oct 14, 2024 • 52

  • Make LoRA Great Again: Boosting LoRA with Adaptive Singular Values and Mixture-of-Experts Optimization Alignment

    Paper • 2502.16894 • Published Feb 24 • 31

  • Ring-lite: Scalable Reasoning via C3PO-Stabilized Reinforcement Learning for LLMs

    Paper • 2506.14731 • Published Jun 17 • 9

  • SlimMoE: Structured Compression of Large MoE Models via Expert Slimming and Distillation

    Paper • 2506.18349 • Published Jun 23 • 13
Upvote
-
  • Collection guide
  • Browse collections
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs