Back to List
Notice:This resource is provided by a third-party author. Please review the code with AI tools or manually before use to ensure security and compatibility.
Cmicrosoft/Tutel

Tutel

Tutel MoE: Optimized Mixture-of-Experts Library, Support GptOss/DeepSeek/Kimi-K2/Qwen3 using FP8/NVFP4/MXFP4

78.7/100
973Forks: 107
View on GitHub
Loading report...

Similar Projects

cactus

85

Low-latency AI engine for mobile devices & wearables

C4.4K

picolm

64

Run a 1-billion parameter LLM on a $10 board with 256MB RAM

C1.3K

esp-ai

74

The simplest and lowest-cost AI integration solution. If you like this project, please give it a Star~ | 最简单、最低成本的AI接入方案。喜欢本项目的话点个 Star 吧~

C772

nanolang

77

A tiny experimental language designed to be targeted by coding LLMs

C570
Back to List