Serverless LLM Serving for Everyone.
A high-throughput and memory-efficient inference and serving engine for LLMs
Machine Learning Engineering Open Book
TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.
Supercharge Your LLM with the Fastest KV Cache Layer