A Framework for Domain-Specific Evaluations

Harvey is expanding our public-facing evaluation work in four key areas: Insights, Research, Approaches, and Context.

Oct 27, 2025

At Harvey, we spend a lot of time thinking about evaluation of large language models (LLMs) on domain-specific applications. Where important for the public conversation about evaluation, we have published our research to further the discussion about how this evaluation could be done, or done better.

Today, we think the most important piece missing from the evaluation discourse is not a research finding or benchmark, but rather a clear, shared understanding of what evaluation of LLMs does and does not tell us. Failing to build a communal understanding of evaluation risks evaluation succumbing to Goodhart’s law: ”When a measure becomes a target, it ceases to be a good measure.” While models will continue to saturate benchmarks, those gains will fail to translate into real value.

At worst, misaligned benchmarks may actively push models towards unhelpful behaviors as showcased by OpenAI’s recent work arguing that current evaluative methods increase hallucinations by penalizing abstention under uncertainty. These risks are particularly acute in domain-specific evaluation, where limited access to expensive domain experts makes sparse or misaligned evaluations all the more likely.

Solving for this requires not just better evaluation, but better understanding of evaluation by both model producers and consumers. To start closing this gap, we are expanding our public-facing evaluation work in four key areas: Insights, Research, Approaches, and Context. In the same way the IRAC framework helps break down legal reasoning, we use this acronym to capture the four pillars of building evaluations that are both rich and accessible. Throughout the rest of this post, we’ll dig deeper into these four pillars and our forthcoming work across each of them.

Insights

How well is AI performing at a particular task at a particular point in time?

Insights are the core of evaluation. An insight is a single number describing how a model performs on a particular task or set of tasks. For example, our main Biglaw Bench (BLB) evaluation describes how well models complete real world legal tasks. It also produces derivative insights on how well those tasks are performed across practice areas and task types.

While these numbers can be reductive, they are essential to building scalable and comprehensible evaluations. Most people do not have the time to routinely engage with the underlying data — that’s our job. Well-aligned insights allow us to aggregate and communicate that data efficiently, creating a useful shorthand for understanding and engaging with evaluations.

Insights will continue to headline most of our work in evaluation. As important as the larger conversation about evaluation is, the most important thing remains that we can identify and describe exactly how our systems are delivering value today and how that value is improving over time. Insights are the broadest way to communicate those critical features and start deeper conversations about what exactly those insights mean and how they can be refined.

Research

How are we evolving our benchmarks to better understand models and agents?

Research represents longer-term work on evaluative methodologies aimed at creating new insights or improving existing insights. Currently, our research focuses extensively on evaluative utility — defining benchmarks that provide direct and meaningful insights about model performance in the real world.

As benchmarks continue to evolve, we think that the essential question becomes building evaluations representing the extremes of model performance. On the one hand, evaluations should establish where models are exceptionally good — these are today’s high value use cases. On the other, evaluations should identify meaningful use cases where models continue to struggle — these establish the continuing frontier for model development.

To support both ends of this spectrum, we will be publishing at least two novel benchmarks. First, our work on Contract Intelligence, which focuses on understanding which contracts models can analyze as well as human experts and how models can express this analysis in different forms. Second, we are supplementing our BLB core benchmark with a BLB Challenge benchmark focused on queries where we see leading models and agents struggle to solve tasks critical to supporting legal work.

Approaches

How do we operationalize evaluations and use them to drive improvements?

In addition to large benchmarks, we run many other evaluation loops to both tell how good our products are doing, and how to make them better. Approaches describe the end-to-end ways we operationalize and digest these various evaluative signals to determine what’s good, what’s better, and where we can invest to make systems their best.

Approaches also describe the nuts and bolts of evaluation, answering questions like:

  1. How do we incorporate lawyers into both producing and understanding our evaluations?
  2. How do we listen to and incorporate client feedback into evaluations while adhering to strict data privacy?
  3. How do we ensure that systems work well across jurisdictions, languages, and practice areas?
  4. How do we convert expert review into automated evaluation systems using LLMs and other signals?

In short, approaches cover the people and practices that turn our research into insights and insights into improvements.

Context

How do we ground our understanding of what evaluation does and does not tell us?

Context ties evaluation together and makes it accessible through plain-language explanations of the concepts, terms, and methodologies behind evaluative methods. Building and sharing context on AI and evaluation is built into Harvey’s DNA. Since day one, we have hired smart lawyers with no AI experience (and smart AI engineers with no legal experience) and taught them these concepts from scratch. Bringing these conversations to the public can help make evaluations both meaningful and actionable.

Take, for example, recent benchmarks of real-world tasks. Mercor claims GPT-5 is the most economically valuable model. OpenAI (the makers of GPT-5) disclaim this, ceding that title to Anthropic’s Claude Opus 4.1. Our own Biglaw Bench gives GPT-5 a modest edge, though the recently released Claude Sonnet 4.5 closes the gap. The leaders also come in at 70%, 47%, and 89% on these respective benchmarks. At face value, these insights are somewhere between meaningless and paradoxical. They certainly do not tell a time-pressured associate which model will help them deliver a great outcome and maybe get a few more hours of sleep.

Context on the means, methods, and measures of these benchmarks can help cut through this confusion. It will also, unfortunately, tell you that none of those benchmarks can give direct guidance to that associate — we’d suggest they use Harvey. Knowing what evaluation currently does and does not tell us is the key to productive conversations on the current and future state of AI. Context is what makes those conversations possible.

What's to Come

As models and agents continue to improve, building a broad coalition on evaluation is necessary to ensure these improvements translate into real-world value. Building that coalition requires not only pushing the frontiers of evaluation, but also ensuring that those frontiers are understood and explorable by everyone. We are excited to start building that community, so that together we can keep building the best AI for legal work.