Loading connector details…
Loading connector details…
Choose a unique username to continue using AgentHotspot
by Yoosu-L • Uncategorized
An MCP server to benchmark Large Language Model APIs measuring throughput and latency metrics.
Measure and analyze the performance of LLM API endpoints.
Benchmark LLM APIs with customizable concurrency and token limits.
Detailed throughput and latency metrics for LLM model evaluation.
This MCP server enables comprehensive benchmarking of LLM APIs by measuring generation throughput, prompt throughput, and Time To First Token (TTFT). It supports flexible deployment options including remote SSE servers and local stdio or SSE transports. The server provides detailed JSON output for in-depth performance analysis and is designed for easy integration with MCP clients.