Honest Pricing for Honest Software.
Lifetime license. Your hardware, your data. Pay once, own forever.
Lifetime license — not a subscription. Your benchmark content stays on your machine. Optional crash diagnostics can be disabled. 30-day satisfaction promise.
What you're reserving.
Full plan detail below. AiBenchLab is launching soon — all purchases open at launch. Founding 100 reservations lock in 50% off before the public sale.
Trial
Kick the tires
- ✓ Built-in llama.cpp server + bundled model
- ✓ 51,000+ model catalog with GPU Fit detection
- ✓ 8-step guided wizard with AI recommendations
- ✓ 2 pre-built suites (Quick Compare + Customer-Facing Chat)
- ✓ 3 models per benchmark session
- ✓ Composite scores, per-domain breakdown, individual test results
- ✓ Watermarked PDF report (1 per session)
- ✓ MBX signed export
- ✓ 14-day full trial, then permanent free limits
Pro
Developers and teams deploying AI to production
Then $399/yr for updates (optional)
- Everything in Trial, plus:
- ✓ All 22 pre-built suites
- ✓ 10 models per session
- ✓ TTFT, TPOT, TPS, E2E Latency metrics
- ✓ Unlimited model comparison
- ✓ Speed vs accuracy scatter plot
- ✓ AI-generated model recommendations
- ✓ Benchmark history
- ✓ Clean PDF single-session reports
- ✓ PDF comparison reports
- ✓ Report section customization (10 sections)
- ✓ JSON export
- ✓ CSV export
- ✓ Company name and logo on reports
- ✓ Tagline and contact info on reports
- ✓ Custom report footer
- ✓ Model fingerprinting (SHA-256)
- ✓ GGUF metadata extraction
- ✓ Model Queue + Scheduled Runs
- ✓ Context Window testing (MRCR)
- ✓ CLI interface
- ✓ 12 months of updates included
Consultant
AI consultants, agencies, freelancers
Then $1,999/yr for updates (optional)
- Everything in Pro, plus:
- ✓ 15 models per session
- ✓ Custom Suite builder with domain weights and pass thresholds
- ✓ Save As New Suite for reusable evaluation templates
- ✓ White-label reports — remove AiBenchLab branding
- ✓ Plugin management
- ✓ REST API for automation and CI/CD
- ✓ MCP Server for AI tool ecosystem integration
- ✓ 3 seats included
- ✓ Priority + direct support
Enterprise
Teams and organizations
Annual contract — everything included
- Everything in Consultant, plus:
- ✓ Site license (per location)
- ✓ Custom test creation — build tests for your internal standards
- ✓ All updates included for duration of contract
- ✓ Dedicated support
- ✓ Custom terms and onboarding
Full feature comparison.
| Feature | Trial | Pro | Consultant | Enterprise |
|---|---|---|---|---|
| Test Suites | ||||
| Pre-built suites available | 2 | All 22 | All 22 | All 22 |
| Add/remove tests from suite | — | ✓ | ✓ | ✓ |
| Create custom suites from scratch | — | — | ✓ | ✓ |
| Save As New Suite | — | — | ✓ | ✓ |
| Benchmarking | ||||
| Models per session | 3 | 10 | 15 | 15 |
| 8-step guided wizard | ✓ | ✓ | ✓ | ✓ |
| AI-powered model recommendations | ✓ | ✓ | ✓ | ✓ |
| Cancel / Abort | ✓ | ✓ | ✓ | ✓ |
| Model Queue | — | ✓ | ✓ | ✓ |
| Scheduled Runs | — | ✓ | ✓ | ✓ |
| Results & Scoring | ||||
| Composite score | ✓ | ✓ | ✓ | ✓ |
| Per-domain breakdown | ✓ | ✓ | ✓ | ✓ |
| Individual test results (expandable) | ✓ | ✓ | ✓ | ✓ |
| TTFT, TPOT, TPS, E2E Latency | — | ✓ | ✓ | ✓ |
| Model comparison (side-by-side) | — | Unlimited | Unlimited | Unlimited |
| Speed vs accuracy scatter plot | — | ✓ | ✓ | ✓ |
| AI-generated recommendations | — | ✓ | ✓ | ✓ |
| Benchmark history | — | ✓ | ✓ | ✓ |
| Reports & Export | ||||
| View results in app | ✓ | ✓ | ✓ | ✓ |
| PDF single-session report | Watermarked | ✓ | ✓ | ✓ |
| Individual test details in report | Redacted | ✓ | ✓ | ✓ |
| PDF comparison report | — | ✓ | ✓ | ✓ |
| JSON export | — | ✓ | ✓ | ✓ |
| CSV export | — | ✓ | ✓ | ✓ |
| MBX signed export | ✓ | ✓ | ✓ | ✓ |
| Batch export (ZIP, CLI/API) | — | ✓ | ✓ | ✓ |
| Report section customization | — | ✓ | ✓ | ✓ |
| Branding | ||||
| Company name & logo on reports | — | ✓ | ✓ | ✓ |
| Tagline and contact info | — | ✓ | ✓ | ✓ |
| Custom report footer | — | ✓ | ✓ | ✓ |
| Remove AiBenchLab branding (white-label) | — | — | ✓ | ✓ |
| Advanced Features | ||||
| Context Window testing (MRCR) | — | ✓ | ✓ | ✓ |
| Model fingerprinting (SHA-256) | — | ✓ | ✓ | ✓ |
| GGUF metadata extraction | — | ✓ | ✓ | ✓ |
| Plugin management | — | — | ✓ | ✓ |
| Custom test creation | — | — | — | ✓ |
| Interfaces | ||||
| GUI Desktop (Tauri) | ✓ | ✓ | ✓ | ✓ |
| CLI interface | — | ✓ | ✓ | ✓ |
| REST API | — | — | ✓ | ✓ |
| MCP Server | — | — | ✓ | ✓ |
| Business & Support | ||||
| Seats | 1 | 1 | 3 | Site license |
| Support level | Community | Priority email | Priority + direct | Dedicated |
| Founding 100 direct channel | — | ✓ | ✓ | ✓ |
| Updates | ||||
| 12 months of updates included | — | ✓ | ✓ | ✓ |
| Annual update fee | — | $399/yr | $1,999/yr | Included |
12 months included. After that, it's optional.
Your license works forever. The annual update subscription is for version upgrades, new test suites, new provider integrations, and security patches.
| Tier | Annual Renewal |
|---|---|
| Pro | $399/yr |
| Consultant | $1,999/yr |
| Enterprise | Included |
The wrong AI model costs more than the right test.
Pro — $999
Pays for itself in a week.
All 22 suites, comparison reports, full export. 998 scoring dimensions across 254 tests and 11 domains on your hardware. Lifetime.
Consultant — $4,999
One client engagement pays for it.
White-label reports look like your own proprietary methodology. The license pays for itself on your first project — everything after that is pure margin.
System Requirements
Minimum
- 8 GB RAM
- 2 GB disk space
- 64-bit OS (Windows 10/11)
- Internet for cloud models
Recommended
- 16+ GB RAM
- NVIDIA GPU with 8+ GB VRAM
- SSD storage
- Windows 11 64-bit
Frequently Asked Questions
You pay once and own that version of AiBenchLab forever. This is not SaaS — there is no server on our end running your benchmarks, no account that gets deactivated. Your data never leaves your machine unless you explicitly use a cloud AI provider. AiBenchLab runs on your hardware, stores data in your local database, and works without an internet connection.
12 months of updates are included with every license purchase. After that, the Optional Annual Update Subscription keeps you current with all version upgrades, new test suites and domains, new provider integrations, and security patches. If you don't renew, nothing breaks — your current version continues to work exactly as it did. If you renew later, you pick up where you left off.
Yes. You pay only the difference between your current tier and the new one.
30-day satisfaction promise. Talk to us within the first 30 days, and if we can't solve it, you get a full refund. Fair enough?
The first 100 people to purchase any paid tier get 50% off their lifetime license. They also get a Founding Member badge, direct access to the developer, and priority on feature requests.
Spots fill in waitlist order. Reserve yours from the pricing page and you'll be notified at launch with your spot confirmed before any public sale opens. Once 100 are claimed, the program closes.
Trial includes Quick Compare and Customer-Facing Chat. Pro unlocks all 22 suites: the full Production category (Agent Readiness, Enterprise Safety, API Reliability, Enterprise Quality Gate, Context MRCR, Multi-Turn Safety), Role-Specific (Coding Assistant, Content & Writing, Reasoning & Analysis, Multimodal), Comparison (Regression Check, Full Benchmark), and App-Specific (OpenClaw Readiness, n8n Workflow Ready, RAG Pipeline, Local Copilot Ready, Creative & Content Studio, Document Analyst, Roleplay & Character, MCP Tool Use Ready).
For local models (Ollama, LM Studio): No. Everything runs locally. Your prompts, responses, and results never leave your machine. For cloud providers (OpenAI, Anthropic, Gemini, Grok, Groq): Test prompts are sent to their API as part of normal API usage. Results are stored locally.
Windows 10/11 (64-bit) at launch. macOS and Linux builds are planned.
For testing cloud models (OpenAI, Anthropic, etc.), no — any machine works. For testing local models, you need whatever hardware those models require (typically a GPU with sufficient VRAM). GPU Fit detection in the Model Catalog helps you find models that fit your hardware.
Because there is no other tool that does what AiBenchLab does. Public benchmarks test on cloud servers under ideal conditions. AiBenchLab is the only professional-grade benchmarking application that runs 998 scoring dimensions across 254 tests and 11 domains on YOUR hardware, with deployment risk scoring, forensic reporting, and deterministic evaluation. You're not comparing this to a $10/month SaaS — you're comparing it to the cost of deploying the wrong model in production.
All providers are available on every tier, including Trial. Local: Built-in llama.cpp server, Ollama, LM Studio, LocalAI. Cloud: OpenAI, Anthropic, Google Gemini, xAI (Grok), Groq. Plus any custom OpenAI-compatible endpoint. That's 10 providers on day one.