{"package": "nm-vllm", "summary": "A high-throughput and memory-efficient inference and serving engine for LLMs", "pypi_url": "https://pypi.org/project/nm-vllm", "piwheels_url": "https://www.piwheels.org/project/nm-vllm", "releases": {"0.6.3.0": {"released": "2024-11-06 13:32:38", "prerelease": false, "yanked": false, "skip_reason": "binary only", "files": {}}, "0.5.3.0": {"released": "2024-09-05 07:02:25", "prerelease": false, "yanked": false, "skip_reason": "binary only", "files": {}}, "0.5.2.0": {"released": "2024-08-12 14:46:43", "prerelease": false, "yanked": false, "skip_reason": "binary only", "files": {}}, "0.5.1.1": {"released": "2024-07-17 17:45:09", "prerelease": false, "yanked": false, "skip_reason": "binary only", "files": {}}, "0.4.0": {"released": "2024-07-11 13:05:26", "prerelease": false, "yanked": false, "skip_reason": "", "files": {}}, "0.2.0": {"released": "2024-04-09 18:30:05", "prerelease": false, "yanked": false, "skip_reason": "", "files": {}}, "0.1.0": {"released": "2024-03-02 02:22:05", "prerelease": false, "yanked": false, "skip_reason": "binary only", "files": {}}}}