User Guide¶
This guide covers everything you need to know to effectively use GenAI Bench for benchmarking LLM endpoints.
What You'll Learn¶
-
Running Benchmarks
Learn how to run benchmarks against various LLM endpoints
-
Multi-Cloud Setup
Configure authentication for AWS, Azure, GCP, OCI, and more
-
Docker Deployment
Run GenAI Bench in containerized environments
-
Excel Reports
Generate comprehensive Excel reports from benchmark results
Common Workflows¶
Basic Benchmarking¶
- Choose your model provider - OpenAI, AWS Bedrock, Azure OpenAI, etc.
- Configure authentication - API keys, IAM roles, or service accounts
- Run the benchmark - Specify task type and parameters
- Analyze results - View real-time dashboard or generate reports
Cross-Cloud Benchmarking¶
Benchmark models from one provider while storing results in another:
# Benchmark OpenAI, store in AWS S3
genai-bench benchmark \
--api-backend openai \
--api-key $OPENAI_KEY \
--upload-results \
--storage-provider aws \
--storage-bucket my-results
Multi-Modal Tasks¶
Support for text, embeddings, and vision tasks:
text-to-text
- Chat and completion taskstext-to-embeddings
- Embedding generationimage-text-to-text
- Vision-language taskstext-to-rerank
- Document reranking
Need Help?¶
- Check the Quick Reference for common commands
- Review Command Guidelines for detailed options
- See Troubleshooting for common issues