Back to List
Notice:This resource is provided by a third-party author. Please review the code with AI tools or manually before use to ensure security and compatibility.
C++lemonade-sdk/lemonade

lemonade

Lemonade helps users discover and run local AI apps by serving optimized LLMs right from their own GPUs and NPUs. Join our discord: https://discord.gg/5xXzkMu8Zk

85.6/100
3.6KForks: 270
View on GitHubHomepage →
Loading report...

Similar Projects

FastFlowLM

82

Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.

C++1.2K

PowerInfer

57

High-speed Large Language Model Serving for Local Deployment

C++9.4K

ZhiLight

65

A highly optimized LLM inference acceleration engine for Llama and its variants.

C++904

distributed-llama

81

Distributed LLM inference. Connect home devices into a powerful cluster to accelerate LLM inference. More devices means faster inference.

C++2.9K
Back to List