llmperf
is a benchmarking tool for stress testing self-hosted LLMs, Azure OpenAI, and OpenAI APIs. It provides efficient concurrent request management, detailed performance analysis, and helps developers optimize inference latency and throughput. With customizable test parameters, it is suitable for various LLM workload evaluation scenarios.
Copy and set up the environment variables from .env.example
.
cp .env.example .env
Copy and set up the environment variables.
./start.sh