replicate/vllm-with-loras
A high-throughput and memory-efficient inference and serving engine for LLMs
Stars: 6Language: Python
Give AlbumentationsX a star on GitHub — it powers this leaderboard
Star on GitHubA high-throughput and memory-efficient inference and serving engine for LLMs