llm-api-benchmark-mcp-server

Runs a throughput benchmark for LLM APIs, measuring generation throughput, prompt throughput, and Time To First Token (TTFT) under various concurrency levels.

Installation

In a virtualenv (see these instructions if you need to create one):

pip3 install llm-api-benchmark-mcp-server

Releases

Version Released Bullseye
Python 3.9
Bookworm
Python 3.11
Files
0.1.4 2025-06-26
0.1.2 2025-06-26
0.1.1 2025-06-26
0.1.0 2025-06-26

Issues with this package?

Page last updated 2025-06-27 15:47:22 UTC