{"package": "vllm-online", "summary": "A high-throughput and memory-efficient inference and serving engine for LLMs", "pypi_url": "https://pypi.org/project/vllm-online", "piwheels_url": "https://www.piwheels.org/project/vllm-online", "releases": {"0.4.2": {"released": "2024-04-29 02:49:29", "prerelease": false, "yanked": false, "skip_reason": "", "files": {}}, "0.4.1": {"released": "2024-04-29 00:17:12", "prerelease": false, "yanked": false, "skip_reason": "", "files": {}}}}