Skip to main content

Benchmarking Domain Intelligence


Share this post
Blazingly Fast LLM Evaluation for In-Context Learning

Large language models are improving rapidly; to date, this improvement has largely been measured via academic benchmarks. These benchmarks, such as MMLU and BIG-Bench, have been adopted by researchers in an attempt to compare models across various dimensions of capability related to general intelligence. However enterprises care about the quality of AI systems in specific domains, which we call domain intelligence. Domain intelligence involves data and tasks that deal with the inner workings of business processes: details, jargon, history, internal practices and workflows, and the like.

Therefore, enterprise practitioners deploying AI in real-world settings need evaluations that directly measure domain intelligence. Without domain-specific evaluations, organizations may overlook models that would excel at their specialized tasks in favor of those that score well on possibly misaligned general benchmarks. We developed the Domain Intelligence Benchmark Suite (DIBS) to help Databricks customers build better AI systems for their specific use cases, and to advance our research on models that can leverage domain intelligence. DIBS measures performance on datasets curated to reflect specialized domain knowledge and common enterprise use cases that traditional academic benchmarks often overlook.

In the remainder of this blog post, we will discuss how current models perform on DIBS in comparison to comparable academic benchmarks. Our key takeaways include:

  1. Models’ rankings across academic benchmarks do not necessarily map to their rankings across industry tasks. We find discrepancies in performance between academic and enterprise rankings, emphasizing the need for domain-specific testing.
  2. There is room for improvement in core capabilities. Some enterprise needs like structured data extraction show clear paths for improvement, while more complex domain-specific tasks require more sophisticated reasoning capabilities.
  3. Developers should choose models based on specific needs. There is no single best model or paradigm. From open-source options to retrieval strategies, different solutions excel in different scenarios.

This underscores the need for developers to test models on their actual use cases and avoid limiting themselves to any single model option.

Introducing our Domain Intelligence Benchmark Suite (DIBS) 

DIBS focuses on three of the most common enterprise use cases surfaced by Databricks customers: 

  1. Data Extraction: Text to JSON 
    • Converting unstructured text (like emails, reports, or contracts) into structured JSON formats that can be easily processed downstream.
  2. Tool Use: Function Calling
    • Enabling LLMs to interact with external tools and APIs by generating properly formatted function calls.
  3. Agent Workflows: Retrieval Augmented Generation (RAG)
    • Enhancing LLM responses by first retrieving relevant information from a company's knowledge base or documents.

We evaluated fourteen popular models across DIBS and three academic benchmarks, spanning enterprise domains in finance, software, and manufacturing. We are expanding our evaluation scope to include legal, data analysis and other verticals, and welcome collaboration opportunities to assess additional industry domains and tasks.

In Table 1, we briefly provide an overview of each task, the benchmark we have been using internally, and academic counterparts if available. Later, in Benchmark Overviews, we discuss these in more detail.

 
Task Category Dataset Name Enterprise or Academic Domain Task Description

Data Extraction:

Text to JSON

Text2JSON Enterprise Misc. Information Given a prompt containing a schema and a few Wikipedia-style paragraphs, extract relevant information into the schema.
Tool Use: Function Calling BFCL-Full Universe Enterprise Function calling Modification of BFCL where, for each query, the model has to select the correct function from the full set of functions present in the BFCL universe.
Tool Use: Function Calling BFCL--Retrieval Enterprise Function calling Modification of BFCL where, for each query, we use text-embedding-3-large to select 10 candidate functions from the full set of functions present in the BFCL universe. The task then becomes to choose the correct function from that set.
Tool Use: Function Calling Nexus Academic APIs Single turn function calling evaluation across 7 APIs of varying difficulty
Tool Use: Function Calling

Berkeley Function Calling Leaderboard

(BFCL)

Academic Function calling See original BFCL blog.

Agent Workflows: RAG

DocsQA Enterprise Software - Databricks Documentation with Code Answer real user questions based on public Databricks documentation web pages.

Agent Workflows: RAG

ManufactQA Enterprise

Manufacturing - Semiconductors -Customer FAQs


 
Given a technical customer query about debugging or product issues, retrieve the most relevant page from a corpus of hundreds of product manuals and datasheets, and construct an answer like a customer support agent.

Agent Workflows: RAG

FinanceBench Enterprise Finance - SEC Filings Perform financial analysis on SEC filings, from Patronus AI

Agent Workflows: RAG

Natural Questions Academic Wikipedia Extractive QA over Wikipedia articles

Table 1. We evaluate the set of models across 9 tasks spanning 3 enterprise task categories: data extraction, tool use, and agent workflows. The three categories we discuss were selected due to their relative frequency in enterprise workloads. Beyond these categories, we are continuing to expand to a broader set of evaluation tasks in collaboration with our customers.

What We Learned Evaluating LLMs on Enterprise Tasks

Academic Benchmarks Obscure Enterprise Performance Gaps

Table showing spread of models
Figure 1.  There is a clear difference in the spread of model performance between academic and enterprise benchmarks. Evaluating models based purely on their academic benchmark performance may lead to subpar model selection. Here, the academic RAG average comprises NQ, while the enterprise RAG average comprises FinanceBench, DocsQA, and ManufactQA. The Academic FC average comprises BFCL and Nexus, while the Enterprise FC average comprises our two modified and more enterprise-oriented variants of BFCL.

In Figure 1, we show a comparison of RAG and function calling (FC) capabilities between the enterprise and academic benchmarks, with average scores plotted for all fourteen models. While the academic RAG average has a larger range (91.14% at the top, and 26.65% at the bottom), we can see that the vast majority of models score between 85% and 90%. The enterprise RAG set of scores has a narrower range, because it has a lower ceiling – this reveals that there is more room to improve in RAG settings than a benchmark like NQ might suggest. 

Figure 1 visually reveals wider performance gaps in enterprise RAG scores, shown by the more dispersed distribution of data points, in contrast to the tighter clustering seen in the academic RAG column. This is most likely because academic benchmarks are based on general domains like Wikipedia, are public, and are several years old - therefore, there is a high probability that retrieval models and LLM providers have already trained on the data. For a customer with private, domain specific data though, the capabilities of the retrieval and LLM models are more accurately measured with a benchmark tailored to their data and use case. A similar effect can be observed, though it is less pronounced, in the function calling setting. 

Structured Extraction (Text2JSON) presents an achievable target

Bar graph comparing model performance on the Text2JSON benchmark
Figure 2.  Many models score very similarly on the Text2JSON task. There remains significant room for improvement; the highest set of scores is ~ 62%.

At a high level, we see that most models have significant room for improvement in prompt-based Text2JSON; we did not evaluate model performance when using structured generation. 

Figure 2 shows that on this task, there are three distinct tiers of model performance:

  • Most closed-source models as well as Llama 3.1 405B and 70B score around just 60%
  • Claude 3.5 Haiku, Llama 3.1 8B and Gemini 1.5 Flash bring up the middle of the pack with scores between 50% and 55%. 
  • The smaller Llama 3.2 models are much worse performers. 

Taken together, this suggests that prompt-based Text2JSON may not be adequate for production use off-the-shelf even from leading model providers. While structured generation options are available, they may impose restrictions on viable JSON schemas and be subject to different data usage stipulations. Fortunately, we have had success fine-tuning models to improve at this capability.  

Other tasks may require more sophisticated capabilities

We also found FinanceBench and Function Calling with Retrieval to be challenging tasks for most models. This is likely because the former requires a model to be proficient with numerical complexity, and the latter requires an ability to ignore distractor information. 

No Single Model Dominates all Tasks

Bar graph showing top 3 performing models for each benchmark
Figure 3.  Overall, GPT-o1-preview, Claude 3.5 Sonnet (New) and GPT-o1-mini are the most consistent top performers. These three models also perform the best on data extraction and RAG tasks. The Gemini 1.5 models are the only models with a context window large enough to perform the full universe BFCL variant. Llama 3.1 405B and 70B are surprisingly strong at the BFCL with retrieval variant.

Our evaluation results do not support the claim that any one model is strictly superior to the rest. Figure 3 demonstrates that the most consistently high-performing models were o1-preview, Claude Sonnet 3.5 (New), and o1-mini, achieving top scores in 5, 4, and 3 out of the 6 enterprise benchmark tasks respectively. These same three models were overall the best performers for data extraction and RAG tasks. However, only Gemini models currently have the context length necessary to perform the function calling task over all possible functions. Meanwhile, Llama 3.1 405B outperformed all other models on the function calling as retrieval task. 

Small models were surprisingly strong performers: they mostly performed similarly to their larger counterparts, and sometimes significantly outperformed them. The only notable degradation was between o1-preview and o1-mini on the FinanceBench task. This is interesting given that, as we can see in Figure 3, o1-mini outperforms o1-preview on the other two enterprise RAG tasks. This underscores the task-dependent nature of model selection.  

Open Source vs. Closed Source Models

Comparison chart between llama model performance and GPT-4o
Figure 4. Llama 3.1 405B performs similarly to GPT-4o on many tasks. We can see that performance quality somewhat falls off as model size decreases, although Llama 3.2 3B punches above its weight on some of the academic benchmarks we evaluate.

We evaluated five different Llama models, each at a different size. In Figure 4, we plot the scores of each of these models on each of our benchmarks against GPT-4o’s scores for comparison. We find that Llama 3.1 405B and Llama 3.1 70B perform extremely competitively on Text2JSON and Function Calling tasks as compared to closed-source models, surpassing or performing similarly to GPT 4o. However, the gap between these model classes is more pronounced on RAG tasks. 

Additionally, we note that Llama 3.1 and 3.2 series of models show diminishing returns regarding model scale and performance. The performance gap between Llama 3.1 405B and Llama 3.1 70B is negligible on the Text2JSON task, and significantly smaller on every other task than Llama 3.1 8B. However, we observe that Llama 3.2 3B outperforms Llama 3.1 8B on the function calling with retrieval task (BFCL-Retrieval in Figure 4).  

This implies two things. First, open-source models are off-the-shelf viable for at least two high-frequency enterprise use cases. Second, there is room to improve these models’ ability to leverage retrieved information. 

To further investigate this, we compared how much better each model would perform on ManufactQA under a closed book setting vs. a default RAG setting. In a closed book setting, models are asked to answer the queries without any given context - which measures a model’s pretrained knowledge. In the default RAG setting, the LLM is provided with the top 10 documents retrieved by OpenAI’s text-embedding-3-large, which had a recall@10 of 81.97%.  This represents the most realistic configuration in a RAG system. We then calculated the relative error reduction between the rag and closed book settings. 

Based on Figure 5, we observe that the GPT-o1-mini (surprisingly!) and Claude-3.5 Sonnet are able to leverage retrieved context the most, followed by GPT-o1-preview and Claude 3.5 Haiku. The open source Llama models and Gemini models all trail behind, suggesting that these models have more room to improve in leveraging domain specific context for RAG.

Bar graph comparing model performance on the ManufactQA benchmark
Figure 5. Using the ManufactQA dataset, we measured how much models improved when given access to relevant documents compared to relying solely on their training knowledge. This improvement was calculated as the relative gain in answer correctness: (RAG accuracy - closed-book accuracy) / closed-book accuracy. In closed-book testing, models answered questions using only their trained knowledge, while in the RAG setting, they had access to the 10 most relevant documents retrieved by text-embedding-3-large, simulating real-world RAG deployments. The results revealed a notable pattern: GPT-4 and Claude 3.5 models showed the strongest ability to leverage retrieved context on ManufactQA, while Llama and Gemini models demonstrated comparatively smaller improvements from the additional context.

For function calling at scale, high quality retrieval may be more valuable than larger context windows.

Bar graph on Gemini model performance on BFCL benchmarks
Figure 6. The Gemini 1.5 models outperform when they have access to a retriever as compared to when they have all possible functions available to them. No other model had a context window long enough to be able to perform the full-universe version of BFCL.

Our function calling evaluations show something interesting: just because a model can fit an entire set of functions into its context window does not mean that it should. The only models capable of doing this at this time are Gemini 1.5 Flash and Gemini 1.5 Pro; as Figure 6 displays, those models perform better on the function calling with retrieval variant, where a retriever selects a subset of the full set of functions relevant to the query. The improvement in performance was more remarkable for Gemini 1.5 Flash (~11% improvement) than for Gemini 1.5 Pro (~2.5%). This improvement likely stems from the reality that a well-tuned retriever can increase the likelihood that the correct function is in the context while vastly reducing the number of distractor functions present. Furthermore, we have previously seen that models may struggle with long-context tasks for a variety of reasons.

Benchmark Overviews

Table of benchmark performance numbers for a range of models
Figure 7.   We evaluated fourteen open and closed-source models across six enterprise benchmarks and three analogous academic benchmarks. This heat map (where red is the lowest possible score of 0%, with a deep green at the highest possible score of 100%) shows that our enterprise benchmarks reveal both a wide spread in model performance and that there is significant room for improvement over academic benchmarks.

Having outlined DIBS's structure and key findings, we present a comprehensive summary of fourteen open and closed-source models’ performance across our enterprise and academic benchmarks in Figure 7. Below, we provide detailed descriptions of each benchmark in the remainder of this section.

Data Extraction: Text to JSON 

In today's data-driven landscape, the ability to transform vast amounts of unstructured data into actionable information has become increasingly valuable. A key challenge many enterprises face is building unstructured data to structured data pipelines, either as standalone pipelines or as part of a larger system. 

One common variant we have seen in the field is converting unstructured text – often a large corpus of documents – to JSON. While this task shares similarities with traditional entity extraction and named entity recognition, it goes further – often requiring a sophisticated blend of open-ended extraction, summarization, and synthesis capabilities. 

No open-source academic benchmark sufficiently captures this complexity; we therefore procured human-written examples and created a custom Text2JSON benchmark. The examples we procured involve extracting and summarizing information from passages into a specified JSON schema. We also evaluate multi-turn capabilities, e.g. editing existing JSON outputs to incorporate additional fields and information. To ensure our benchmark reflects actual business needs and provides a relevant assessment of extraction capabilities, we used the same evaluation strategies as our customers. 

Tool Use: Function Calling

Tool use capabilities enable LLMs to act as part of a larger compound AI system. We have seen sustained enterprise interest in function calling as a tool, and we previously wrote about how to effectively evaluate function calling capabilities

Recently, organizations have taken to tool calling at a much larger scale. While academic evaluations typically test models with small function sets—often ten or fewer options—real-world applications frequently involve hundreds or thousands of available functions. In practice, this means enterprise function calling is similar to needle-in-a-haystack test, with many distractor functions present during any given query. 

To better reflect these enterprise scenarios, we've adapted the established BFCL academic benchmark to evaluate both function calling capabilities and the role of retrieval at scale. In its original version, the BFCL benchmark requires a model to choose one or fewer functions from a predefined set of four functions. We built on top of our previous modification of the benchmark to create two variants: one which requires the model to choose from the full set of functions that exist in BFCL for each query, and one which leverages a retriever to identify ten functions that are the most likely to be relevant.

Agent Workflows: Retrieval-Augmented Generation

RAG makes it possible for LLMs to interact with proprietary documents, augmenting existing LLMs with domain intelligence. In our experience, RAG is one of the most popular ways to customize LLMs in practice. RAG systems are also critical for enterprise agents, because any such agent must learn to operate within the context of the particular organization in which it is being deployed. 

While the differences between industry and academic datasets are nuanced, their implications for RAG system design are substantial. Design choices that appear optimal based on academic benchmarks may prove suboptimal when applied to real-world industry data. This means that architects of industrial RAG systems must carefully validate their design decisions against their specific use case, rather than relying solely on academic performance metrics.

Natural Questions remains a popular academic benchmark even as others, such as HotpotQA have fallen out of favor. Both of these datasets deal with Wikipedia-based question answering. In practice, LLMs have indexed much of this information already. For more realistic enterprise settings, we use FinanceBench and DocsQA – as discussed in our earlier explorations on long context RAG – as well as ManufactQA, a synthetic RAG dataset simulating technical customer support interactions with product manuals, designed for manufacturing companies' use cases.

Conclusion

To determine whether academic benchmarks could sufficiently inform tasks relating to domain intelligence, we evaluated a total of fourteen models across nine tasks. We developed a domain intelligence benchmark suite comprising six enterprise benchmarks that represent: data extraction (text to JSON), tool use (function calling), and agentic workflows (RAG). We selected models to evaluate based on customer interest in using them for their AI/ML needs; we additionally evaluated the Llama 3.2 models for more datapoints on the effects of model size. 

Our findings show that relying on academic benchmarks to make decisions about enterprise tasks may be insufficient. These benchmarks are overly saturated – hiding true model capabilities – and somewhat misaligned with enterprise needs. Furthermore, the field of models is muddied: there are several models that are generally strong performers, and models that are unexpectedly capable at specific tasks. Finally, academic benchmark performance may lead one to believe that models are sufficiently capable; in reality, there may still be room for improvement towards being production-workload ready. 

At Databricks, we are continuing to help our customers by investing resources into more comprehensive enterprise benchmarking systems, and towards developing sophisticated approaches to domain expertise. As part of this, we are actively working with companies to ensure we capture a broad spectrum of enterprise-relevant needs, and welcome collaborations. If you are a company looking to create domain-specific agentic evaluations, please take a look at our Agent Evaluation Framework. If you are a researcher interested in these efforts, consider applying to work with us.