VibesMark / README.md
atin121's picture
Added OpenRouter + env file; basic funcitonality working
67a3b3a
|
raw
history blame
1.18 kB

Vibes Benchmark v0.1

A tool for benchmarking different AI models by comparing their responses to custom questions.

Prerequisites

Setup

  1. Clone the repository:

    git clone [repository-url]
    cd vibes-benchmark
    
  2. Install dependencies:

    pip install -r requirements.txt
    
  3. Configure environment variables:

    cp .env.example .env
    

    Then edit .env and add your OpenRouter API key

Usage

  1. Prepare a text file with your questions (one per line)
  2. Run the application:
    python app.py
    
  3. Upload your questions file through the web interface
  4. Click "Run Benchmark" to start comparing model responses

Features

  • Compare responses from different AI models side by side
  • Supports up to 10 questions per benchmark
  • Randomly selects different models for comparison
  • Real-time response generation

Supported Models

  • Claude 3 Opus
  • Claude 3 Sonnet
  • Gemini Pro
  • Mistral Medium
  • Claude 2.1
  • GPT-4 Turbo
  • GPT-3.5 Turbo

License

[Your chosen license]

Run it with python app.py