vllm-rocm

A high-throughput and memory-efficient inference and serving engine for LLMs with AMD GPU support

Installation

In a virtualenv (see these instructions if you need to create one):

pip3 install vllm-rocm

Releases

Version Released Bullseye
Python 3.9
Bookworm
Python 3.11
Files
0.6.3 2024-10-15

Issues with this package?

Page last updated 2024-10-15 17:23:47 UTC