{"package": "evaluate-attacks", "summary": "A package for evaluating adversarial attacks on deep learning models", "pypi_url": "https://pypi.org/project/evaluate-attacks", "piwheels_url": "https://www.piwheels.org/project/evaluate-attacks", "releases": {"0.1.1": {"released": "2024-07-23 08:17:14", "prerelease": false, "yanked": false, "skip_reason": "", "files": {"evaluate_attacks-0.1.1-py3-none-any.whl": {"file_url": "https://archive1.piwheels.org/simple/evaluate-attacks/evaluate_attacks-0.1.1-py3-none-any.whl", "filehash": "dd92a1064778db6cd733da0b052b7cff7de3ebff8bb4c62b3f7286347d4d880e", "filesize": 2586, "builder_abi": "cp311", "file_abi_tag": "none", "platform": "any", "requires_python": null, "apt_dependencies": []}}}}}