Hugging Face
Models
Datasets
Spaces
Posts
Docs
Solutions
Pricing
Log In
Sign Up
WesPro
/
State-of-the-MoE_RP-2x7B
like
1
Text Generation
Transformers
Safetensors
mixtral
text-generation-inference
Inference Endpoints
Model card
Files
Files and versions
Community
1
Train
Deploy
Use this model
No model card
New: Create and edit this model card directly on the website!
Contribute a Model Card
Downloads last month
7
Safetensors
Model size
12.9B params
Tensor type
BF16
·
Inference Examples
Text Generation
Inference API (serverless) is not available, repository is disabled.
Model tree for
WesPro/State-of-the-MoE_RP-2x7B
Quantizations
3 models