We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.
You must be logged in to block users.
Contact GitHub support about this user’s behavior. Learn more about reporting abuse.
Forked from triton-inference-server/fastertransformer_backend
Python
Forked from microsoft/onnxruntime
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
C++
A high-throughput and memory-efficient inference and serving engine for LLMs
Python 77.9k 16k
A blazing fast inference solution for text embeddings models
Rust 4.7k 384
Transformers-compatible library for applying various compression algorithms to LLMs for optimized deployment with vLLM
Python 3.1k 490
There was an error while loading. Please reload this page.