🗻
Highlights
Stars
LLM Serving
LLM 서빙 최적화 관련 오픈소스
4 repositories
TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. Tensor…
SGLang is a fast serving framework for large language models and multi-modality models.
A high-throughput and memory-efficient inference and serving engine for LLMs