Back to List
Notice:This resource is provided by a third-party author. Please review the code with AI tools or manually before use to ensure security and compatibility.
Cmicrosoft/Tutel

Tutel

Tutel MoE: Optimized Mixture-of-Experts Library, Support GptOss/DeepSeek/Kimi-K2/Qwen3 using FP8/NVFP4/MXFP4

75.4/100
986Forks: 108
View on GitHub
Loading report...

Similar Projects

cactus

85

Low-latency AI engine for mobile devices & wearables

C4.7K

picolm

55

Run a 1-billion parameter LLM on a $10 board with 256MB RAM

C1.5K

esp-ai

69

The simplest and lowest-cost AI integration solution. If you like this project, please give it a Star~ | 最简单、最低成本的AI接入方案。喜欢本项目的话点个 Star 吧~

C821

nanolang

78

A tiny experimental language designed to be targeted by coding LLMs

C591
Back to List