Tech Corner December 12, 2023

Using LLMs confidently in financial services workflows

by Elizabeth Matson

Website LLM Traceability

Introducing tools powered by large language models (LLMs) into your workflows presents a new opportunity to dive deeper into your data. The capability to instantly generate detailed summaries of quarterly reports or automate the review of lengthy Confidential Information Memos can be a game changer for your team’s operational efficiency and ability to service more clients.

Conversations about LLMs and generative AI often touch on approaches to maintain data privacy and security standards, particularly in financial services and for enterprise use cases. Another equally important topic is accuracy and traceability. LLMs can occasionally generate incorrect responses, or “hallucinations,” and it’s important that any product using LLMs incorporate layered strategies to reduce hallucinations. If you’re integrating LLMs into your critical workflows, you need to ensure the information is correct and can stand up to regulatory scrutiny. Being confident in the data output from an LLM-powered product is essential.

Taking the time to manually read every document to double-check every answer’s accuracy defeats the purpose of using LLMs in your workflow. So how do you trust that the responses you’re getting are correct? Data scientists researching LLMs and training models are exploring multiple paths to build confidence in LLM outputs:

  • Model explainability: Mechanisms used to explain how a model reasoned over and arrived at an answer, including analyzing the underlying weights and features.
  • Confidence scoring: Scores assigned and provided alongside the model’s response to inform the user whether the response can be taken at face value or requires more interrogation.
  • Attribution: Providing the source for the response—where in a document (or collection of documents) the LLM found an answer.

How Alkymi is centering traceability in our LLM product development

In building our LLM-powered features and tools, our focus is on the needs of our customers, who are primarily in financial services. Data extracted from documents using LLMs needs to be accurate, and it also needs to be traceable. Every piece of actionable data needs to be verified and able to be traced back to the source document, automatically.

Model explainability is useful but doesn’t provide traceability. Similarly, confidence scores provide information to help you gauge your trust in the response, but they’re still not easily verifiable. Attribution is a natural fit to provide traceability for financial services use cases, since it provides a direct link between the LLM's response and the source in the document.

Providing a linked source, or multiple sources, alongside the LLM’s answer to a user’s question allows them to easily find the supporting data for a response without needing to find it independently. Answers can be verified as needed without extensive manual review, whether to confirm the accuracy of the tool and build confidence in its responses or for auditing purposes. Being able to instantly highlight where an answer came from in a long document allows users to validate the output in real-time and provides a clear audit trail.

For example, when reviewing information on a potential investment, you might want to quickly find and synthesize a company’s projected financials into a short summary you can share with your team. Alkymi’s LLM-powered tools can generate that answer and then highlight exactly where in the source document the response originated from.

Traceability example in the Alkymi platform
An example of Alkymi's Traceability functionality in our Document Chat interface, where a source is highlighted in the document, providing attribution for the answer to the user's question.


If you’re looking to introduce any tool utilizing large language models to be useful in financial services, being able to trust its output should be non-negotiable. Our natural instinct is to double-check where the data came from and match it against the original document. An intentionally built tool for financial services data should do that for you, so validating even generative AI can happen in an instant.

Read more about the theory behind our approach in our Data Science Room.

More from the blog

February 22, 2024

Quarterly Statements Rule: 3 ways to ensure compliance

by Elizabeth Matson

The transition window for the SEC's new Private Fund Regulations is well underway. AI-powered workflows will help you stay compliant and improve efficiency.

January 18, 2024

Speed up CIM review with LLMs

by Elizabeth Matson

Reviewing a stack of CIMs can be painful, but it doesn’t have to be. Here's how we redesigned a CIM review workflow using large language models.

December 20, 2023

2023: A year in review

by Harald Collet

2023 was a monumental year for Alkymi. Take a look at our top data points from the past year.