Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

cloudyu
/
Yi-34Bx2-MoE-60B

Text Generation
Transformers
Safetensors
mixtral
yi
Mixture of Experts
conversational
text-generation-inference
Model card Files Files and versions Community
17
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

[AUTOMATED] Model Memory Requirements

#16 opened about 1 year ago by
model-sizer-bot

4x version

1
#15 opened about 1 year ago by
ehartford

Adding Evaluation Results

#14 opened about 1 year ago by
leaderboard-pr-bot

What's the model architecture

#13 opened about 1 year ago by
JamesShao

base or chat model?

#12 opened over 1 year ago by
horaceai

I am a newbie, how to use the existing open source LLM to train MoE. Thank you

#11 opened over 1 year ago by
EEEmpty

vllm

2
#10 opened over 1 year ago by
regzhang

Quantization Please

👍 1
1
#9 opened over 1 year ago by
bingw5

How many GPU memories that the MoE module needs?

2
#8 opened over 1 year ago by
Jazzlee

Multi-langua?

1
#7 opened over 1 year ago by
oFDz

Perfect MoE's my write up, and help to you for making MoE's

👍 3
#6 opened over 1 year ago by
rombodawg

Add MOE (mixture of experts) tag

👍 1
#5 opened over 1 year ago by
davanstrien

What are the merging parameters?

👍 2
3
#4 opened over 1 year ago by
rombodawg

is this base model or sft model?

1
#3 opened over 1 year ago by
lucasjin

Can VLLM be used for inference acceleration?

2
#2 opened over 1 year ago by
obtion

You are all three top spots on the leaderboard

🤗 1
#1 opened over 1 year ago by
dillfrescott
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs