Most AI teams spend 90% of their time on the model, 10% on evaluation. Production teams flip that. I built a tool to make evaluation fast enough to actually happen.
The evaluation problem
Here's the gap I see in most AI projects: teams invest heavily in model selection, infrastructure, and prompt iteration — then ship without a systematic way to know if the prompts actually work. "Looks good to me" is not evaluation.
The symptom is a prompt that passes vibes testing but fails on edge cases users actually hit. Or worse: a prompt that works in staging and breaks silently in production because someone changed the system prompt without running it through the same test cases.
Evaluation solves this. Define what good output looks like before you write the prompt. Run every prompt change against the same test cases. Score automatically. Iterate on the score, not on vibes.
What AI Eval Lab does
AI Eval Lab is a browser-based prompt evaluation tool. You:
- Write a prompt template with a
{{input}}placeholder - Define test cases — each has an input and expected keywords in the output
- Click "Run Evaluation" — all cases run against Claude simultaneously
- Each case scores 0–100% based on keyword match
- Overall score shown across all cases
- "Fix My Prompt" streams improvement suggestions based on failing cases
Why keyword scoring, not semantic similarity
I made a deliberate choice to use keyword matching (does the output contain these strings?) instead of semantic similarity (does this embedding vector match?) for a specific reason: transparency.
Semantic similarity requires you to trust a black-box score. "Your output has 0.87 cosine similarity to the expected output" — OK, but what does that mean? What would make it 0.95? It's hard to act on.
Keyword matching is deterministic and interpretable. "Your output should contain 'wavelength' and 'scattering' — it contains 'scattering' but not 'wavelength'" tells you exactly what to fix. The score is a lagging indicator; the keyword analysis is the signal.
For the kinds of tasks prompt evaluation typically covers — factual recall, structured output, specific instructions — keyword presence is actually a better proxy than semantic similarity. You want the model to say "Maillard reaction," not just something semantically similar.
The "Fix My Prompt" feature
When a prompt underperforms, "Fix My Prompt" sends the original prompt, all test cases, and the actual vs expected outputs to Claude with a system prompt asking for specific, actionable improvements.
The meta-prompt:
You are a prompt engineering expert. Given a prompt template and its
evaluation results, suggest specific improvements.
Focus on:
- Why certain test cases are failing
- What to add, remove, or clarify in the prompt
- Concrete rewrites, not abstract advice
Be specific. Show before/after examples where helpful.The output streams in real time and is often genuinely useful — it identifies patterns in failures that you might not notice looking at individual cases.
Three pre-built examples
The app ships with three working examples to make evaluation immediately tangible:
Explain Rayleigh Scattering — Tests whether the model covers wavelength-dependent scattering, blue light preference, and nitrogen/oxygen interaction. Zero-shot prompts usually miss two of three.
Paris Facts — Tests factual recall about Paris: Eiffel Tower (1889), capital of France, Louvre. Simple but good for demonstrating keyword scoring.
ML Pattern Recognition — Tests technical understanding: overfitting, regularization, cross-validation. Higher bar, more useful for testing specialized prompts.
The evaluation culture argument
The most important thing about evaluation isn't the tooling — it's the discipline. Teams that write test cases before writing prompts produce better AI systems. Not because the tool is magic, but because writing test cases forces you to specify what success looks like.
This is the TDD parallel for AI. You don't write tests after you write code and expect them to be rigorous — you get confirmation bias, testing what you know works. Evaluation written after prompt iteration has the same problem.
The discipline is: before you write the prompt, write three test cases. For each case, write what keywords a good answer must contain. Now you have a definition of success that doesn't depend on your opinion of today's output.
Evaluation is the thing that separates "AI that works in demos" from "AI that works in production." It's not the interesting part. It's the important part.
Connection to Prompt Lab
I think of Prompt Lab and Eval Lab as a two-tool workflow:
- Prompt Lab — exploration. Run the same question through four techniques, find what works, develop intuition.
- Eval Lab — validation. Lock in test cases, iterate on the score, catch regressions before they ship.
Start with Prompt Lab when you're figuring out what prompting approach to use. Graduate to Eval Lab when you need to know if changes made things better or worse. They address different stages of the same workflow.