How Marsh McLennan Saved Over 1 Million Hours of Team Time with Agentic AI Powered by SLMs

March 12, 2025 · 3 min read
Predibase-MarshMcLennan-CaseStudy
Will Van Eaton
Will Van Eaton

Large Language Models (LLMs) have the potential to create greater efficiency, productivity, and savings for enterprises, but getting them to perform accurately for specific use cases remains a challenge. While these models possess vast general knowledge, they often need refinement to truly excel in specialized contexts. Agentic AI, or AI agents, can bridge the gap between theoretical capabilities and practical utility by performing tasks and solving problems specific to an organization's needs.

Using Predibase's fine-tuning capabilities for a couple of specific use cases, global professional services firm  Marsh McLennan launched LenAI–a powerful AI assistant–that enables Marsh McLennan employees to leverage institutional knowledge and provide leading industry expertise to customers quickly and accurately. Their journey demonstrates how thoughtful model customization can bridge the gap between AI's theoretical capabilities and real-world impact.

Background

Marsh McLennan is the world’s leading professional services firm in the areas of risk, strategy and people, with more than 90,000 employees advising clients in 130 countries. Their innovative AI assistant, LenAI, provides Marsh McLennan employees with instant access to accumulated expertise to find the right resources and answers to deliver to customers quickly. 

Marsh McLennan’s Chief Information and Operations Officer Paul Beswick leads 5,000 technologists. Under his leadership, the company adopted a forward-looking approach to generative AI, betting on its potential value before many organizations recognized its enterprise-grade applications. According to Paul, the company decided to move quickly and proactively:

“Our journey with generative AI started before it became a mainstream tool,” Paul explains. “By early 2023, we were making secure APIs available to our teams. By mid-year, we put an LLM-based assistant into pilot, and soon after that, LenAI was launched to our entire global workforce. Today, we handle around 20 million requests annually. It’s become something nearly everyone uses, with upwards of 90,000 people tapping into it.”

Limited intent recognition hampers effectiveness

LenAI, which was initially built on GPT 3.5, started as a chat client that was available to answer team member questions. After about a month, they expanded it to connect many of their internal infrastructures, including platforms like Microsoft SharePoint, internal news feeds, media assets, and specialized company-specific tools. By uniting many sources of truth into one single chat interface, their experts were better equipped to support clients. 

The base GPT 3.5 model occasionally faced difficulties with intent recognition—a crucial capability for any agentic AI–sometimes misinterpreting the correct tool based on a user's query. For example, when users requested simple tasks like drafting emails, the model would return documents containing the keyword ‘email’ rather than launching the email client. This semantic gap between user intent and system response highlighted the need for fine-tuning.

LenAI Genai Marsh McLennan Screen

Fine-tuning LLMs for better results

Marsh McLennan realized that to reach the next level of accuracy and drive task automation with higher precision, they would need to fine-tune their models.  To do so, Marsh McLennan worked with  Predibase, which leverages techniques like low-rank adaptation (LoRA).

“With Predibase, the complexity dropped,” says Paul. “This unlocked a new wave of automation use cases.”

7% increased accuracy over 4o-mini

Fine-tuning not only ensured data sovereignty but also improved the accuracy of their fine-tuned Llama-3.1-8B-instruct model compared to GPT 3.5 and 4o-mini, a notable 10-12% increase over 3.5 and a 7% improvement over 4o-mini. 

Furthermore, fine-tuning with Turbo LoRA–Predibase’s innovation that accelerates SLM throughput while improving accuracy–and deploying on a GPU that supports FP8 quantization enabled LenAI to reduce round trip request time below 4 seconds, something previously impossible with GPT 3.5.

The team recently reached an important milestone. Since launching in December of 2023, LenAI has supported over 25 million cumulative queries across the organization. This translated to a massive increase in productivity, saving at least 1 million hours of team time in its first year alone.

“Since then, we’ve put the power into our colleagues' hands by allowing them to experiment with the same technology that LenAI is using,” said Paul. “We’re offering it to any development team across the organization who wants to experiment with it for a new use case, meaning that the success we’ve seen so far will only grow over time.”

Conclusion

Any organization's backbone is its people and the knowledge and years of experience they draw upon to improve their clients’ businesses. For Marsh McLennan, while not fine-tuning the LLM, has found specific use cases where a fine-tuned AI model powered by Predibase helped them empower their employees with the accurate information they needed to do the job efficiently every day. 

Building on this success, Marsh McLennan continues to expand LenAI’s capabilities. By democratizing access to LenAI and fine-tuning technology across development teams, the company fosters innovation and radically transforms enterprise knowledge management.

What could you achieve with Predibase’s fine-tuning capabilities? Try it for yourself with our 30-day free trial.

Related Articles

Join Our Community!