Skip to main content
Prompts are the instructions that drive your agent’s behavior. Small changes in wording can dramatically affect output quality, but without tracking, you have no way to know which version works best — or even which version is running in production. ZeroEval Prompts gives you version control for prompts with a single function call. Every change is tracked, every completion is linked to the exact prompt version that produced it, and you can deploy optimized versions without touching code.

Why track prompts

  • Version history — every prompt change creates a new version you can compare and roll back to
  • Production visibility — see exactly which prompt version is running, how often it’s called, and what it produces
  • Feedback loop — attach thumbs-up/down feedback to completions, then use it to optimize prompts and evaluate models
  • One-click deployments — push a winning prompt or model to production without redeploying your app

How it works

1

Replace hardcoded prompts

Swap string literals for ze.prompt() calls. Your existing prompt text becomes the fallback content.
2

Versions are created automatically

Each unique prompt string creates a tracked version. Changes in your code produce new versions without any extra work.
3

Completions are linked to versions

When your LLM integration fires, ZeroEval links each completion to the exact prompt version and model that produced it.
4

Optimize from production data

Review completions, submit feedback, and generate improved prompt variants — all from real traffic.

Get started

Python

ze.prompt() and ze.get_prompt() for Python applications

TypeScript

ze.prompt() for TypeScript and JavaScript applications