Back to List
Notice:This resource is provided by a third-party author. Please review the code with AI tools or manually before use to ensure security and compatibility.
Pythonpredibase/lorax

lorax

Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs

85.7/100
3.8KForks: 312
View on GitHubHomepage →
Loading report...

Similar Projects

vllm

93

A high-throughput and memory-efficient inference and serving engine for LLMs

Python80.1K

OpenLLM

89

Run any open-source LLMs, such as DeepSeek and Llama, as OpenAI compatible API endpoint in the cloud.

Python12.3K

peft

91

🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.

Python21.1K

BentoML

88

The easiest way to serve AI apps and models - Build Model Inference APIs, Job queues, LLM apps, Multi-model pipelines, and more!

Python8.6K
Back to List