moe-training
Train Mixture of Experts (MoE) models using DeepSpeed or HuggingFace. Use when training large-scale models with limited compute (5× cost reduction vs dense mode
96/100
Security score
The moe-training skill was audited on Feb 28, 2026 and we found 4 security issues across 1 threat category. Review the findings below before installing.
Categories Tested
Security Issues
low line 514
External URL reference
SourceSKILL.md
| 514 | - **DeepSpeed MoE Tutorial**: https://www.deepspeed.ai/tutorials/mixture-of-experts-nlg/ |
low line 515
External URL reference
SourceSKILL.md
| 515 | - **Mixtral Paper**: https://arxiv.org/abs/2401.04088 |
low line 516
External URL reference
SourceSKILL.md
| 516 | - **Switch Transformers**: https://arxiv.org/abs/2101.03961 |
low line 517
External URL reference
SourceSKILL.md
| 517 | - **HuggingFace MoE Guide**: https://huggingface.co/blog/moe |
Scanned on Feb 28, 2026
View Security Dashboard