How Do You Know if Your Prompt Is Good Enough?
A single question drives the new age of A.I.: how do we measure the quality of what machines generate for us?
The rise of large language models has given rise to a new discipline: prompt engineering. But behind the hype lies a more pressing question. Not every prompt is created equal, and more importantly, not every task requires the same measure of success. The way to evaluate a prompt depends on whether the task is unique or repeated.
Unique tasks: judgement in the loop
For one-off, high-stakes queries, such as writing a legal summary, generating investment insights, or drafting a policy brief, the measure of a good prompt is simple: human evaluation. The user remains in the loop, deciding whether the answer meets the standard. These prompts are closer to creative direction than code: they are bespoke, one-time experiments. Success here is defined by expert judgement, not metrics.
In these cases, iteration is the method. The human adjusts the wording, tests different framings, and evaluates the output directly. There is no need for statistical validation; only a clear-eyed review is required to determine whether the model delivered what was required. Still, there is a technique to build good prompts.
Repeated tasks: measurement at scale
The equation changes when prompts are used repeatedly, such as in customer support replies, automated compliance checks, product descriptions, and financial summaries. In these scenarios, individual judgment no longer scales. What matters is consistency, reliability and statistical performance across hundreds or thousands of runs.
Here, open source and closed models diverge. With open source systems, teams can inspect the model weights directly, fine-tuning and adjusting the architecture for their specific use case. Quality is not just a matter of trial and error; it can be engineered into the system.
Closed systems, however, are black boxes. Here, the only way to judge quality is through statistics. Hundreds of runs must be measured, scored, and compared to establish benchmarks for accuracy, bias, and reliability. Once enough data is collected, fine-tuning can be applied, not by rewriting the model, but by reshaping usage, prompt design and workflow integration.
The strategic divide
The distinction between unique and repeated use cases is more than a technical detail; it is a strategic question for businesses deploying A.I. Unique prompts demand judgment. Repeated prompts demand measurement. Confusing the two can lead to costly mistakes: either wasting time over-optimising a one-off query, or trusting a black-box model without sufficient testing.
The most successful organisations will be those that recognise the difference. They will treat one-off prompts as creative briefs, evaluated by experts. And they will treat repeated prompts as statistical systems, validated through rigorous testing and refined through feedback loops.
Unique prompts demand judgment. Repeated prompts demand measurement
Beyond prompts: towards operational A.I.
The next wave of competitive advantage lies not in writing clever prompts, but in building frameworks to measure and manage them. That means:
Defining whether a task is unique or repeated before deploying a model.
Investing in statistical pipelines for repeated tasks.
Retaining human-in-the-loop evaluation for unique outputs.
Choosing between open source transparency and a closed model scale based on business need.
Prompting is no longer just about asking A.I. the right question. It is about knowing how to judge the answer, and when to measure, when to trust, and when to iterate.


