A benchmarking tool built with the CAMEL framework that compares the performance of various AI models, including OpenAI and Nebius models. This tool measures and visualizes the speed of different models in terms of tokens processed per second.
- Performance benchmarking of multiple AI models
- Visual comparison of model speeds
- Compare any two providers (OpenAI and Nebius models)
- Easy-to-interpret visualizations
- Python 3.11+
- uv - an extremely fast Python package installer and resolver
- Nebius API key and OpenAI API key for the models you want to test
- Clone the repository:
git clone https://github.com/Arindam200/awesome-ai-apps.git
cd starter_ai_agents/camel_ai_starter- Create a virtual environment and install dependencies using
uv:
# Create a virtual environment
uv venv
# Activate the virtual environment
source .venv/bin/activate
# Install dependencies
uv sync- Create a
.envfile and add yourOPENAI_API_KEYandNEBIUS_API_KEY.
OPENAI_API_KEY="your-openai-api-key"
NEBIUS_API_KEY="your-nebius-api-key"Run the main script:
uv run agent.pyThe script will:
- Initialize multiple AI models (OpenAI and Nebius)
- Send a test prompt to each model
- Measure the response time and calculate tokens per second
- Generate a visualization comparing the performance
OpenAI
- GPT-4O Mini
- GPT-4O
Nebius
- Kimi-K2-Instruct
- Qwen3-Coder-480B-A35B-Instruct
- GLM-4.5-Air
