Internal knowledge assistants

Bank staff spend hours searching for policy answers. An SLM-powered knowledge assistant delivers instant, sourced answers from your own documents, on your own infrastructure.

Blog Collection Athour img
Michael Forystek
Co-founder, Growth & Partnerships
shape

Internal knowledge assistants for banking: replacing the policy manual

A compliance officer needs to check the bank's procedure for handling a cross-border wire transfer flagged by the monitoring system. The answer is spread across three internal policy documents, a regulatory guidance note, and an operational manual last updated six months ago. Finding and cross-referencing the relevant sections takes 30-45 minutes. The actual decision takes five.

This scenario repeats thousands of times a day across every financial institution. Staff in operations, compliance, risk, and customer service spend a disproportionate amount of their time searching for information rather than acting on it. The knowledge exists - it is just trapped in documents that are hard to find, hard to search, and hard to interpret under time pressure.

An internal knowledge assistant built on a small language model (SLM) solves this by giving employees instant, natural-language access to the institution's internal policies, procedures, and regulatory guidance - running entirely on the bank's own infrastructure.

Why is internal knowledge management so painful in banking?

The knowledge management challenge in banking is not a technology problem in the traditional sense. Most institutions have document management systems, intranets, and search tools. The problem is that these systems are designed for storage, not retrieval under pressure.

Documents are scattered and siloed

Policy documents live in SharePoint. Regulatory guidance lives in a compliance portal. Operational manuals live in a separate documentation system. Product terms live somewhere else entirely. When a staff member needs an answer that spans multiple sources, they must know where to look, which document is current, and how to reconcile overlapping or conflicting information.

Search is keyword-based, not meaning-based

Traditional search tools match keywords, not intent. Searching for "cross-border wire transfer limits" might return dozens of results - some relevant, some outdated, some from a different department's context. The employee must scan each result and determine which one applies to their specific situation. This is slow and error-prone, especially for newer staff who lack institutional knowledge.

Institutional knowledge is concentrated in experienced staff

In many banks, the fastest way to get an accurate answer to a complex operational question is to ask a senior colleague who has been with the institution for years. This creates a bottleneck - the most experienced people spend a significant portion of their day answering questions from others - and a risk, because that knowledge walks out the door when people leave or retire.

The cost is invisible but significant

Unlike an API bill or a hardware purchase, the cost of poor knowledge management does not appear on a line item. It shows up as longer handling times, inconsistent decisions, repeated training, compliance errors, and employee frustration. These costs are real but rarely measured, which means they are rarely addressed with the urgency they deserve.

What is an internal knowledge assistant?

An internal knowledge assistant is an AI system that employees can query in natural language - just like asking a colleague - and receive accurate, sourced answers drawn from the institution's own documents. Instead of searching through folders and reading through pages, the employee asks a question and gets a direct answer with a reference to the source document.

The system is built on two components: a retrieval-augmented generation (RAG) pipeline that finds the relevant documents, and a small language model that reads the retrieved content and generates a clear, contextual answer.

How it works in practice

An operations analyst types: "What is our current procedure for handling a SAR filing when the customer is a PEP?"

The RAG layer searches the institution's document base - compliance manuals, regulatory guidance, internal procedures - and retrieves the most relevant sections. The SLM reads the retrieved content and generates a concise answer: the specific steps to follow, the escalation path, the documentation requirements, and the relevant policy reference.

The analyst gets their answer in seconds rather than minutes. The answer includes a citation to the source document, so they can verify it and reference it in their records. The entire interaction happens on the bank's own infrastructure - no data leaves the environment.

Why use a small language model instead of a general-purpose LLM?

Financial institutions evaluating AI for internal knowledge retrieval face the same architectural choice as any other AI use case: use a third-party LLM API or deploy a purpose-built model on-premise.

Data sensitivity

Internal policy documents, compliance procedures, and operational manuals contain sensitive institutional information. Sending these documents - or questions about them - to a third-party API creates data exposure. An on-premise SLM keeps everything within the institution's infrastructure, which is especially important for queries that reference specific customer cases, ongoing investigations, or unreleased regulatory interpretations.

Domain accuracy

A general-purpose LLM will attempt to answer a question about your bank's internal wire transfer procedure using its general knowledge of banking. An SLM fine-tuned on your institution's specific language and conventions understands the terminology, abbreviations, and operational context unique to your organization. "CAT-2 escalation" means something specific in your institution that a general model cannot know.

Response consistency

When multiple employees ask the same question, they should get the same answer. A general-purpose LLM may generate different phrasings, different levels of detail, or even different interpretations each time. A fine-tuned SLM produces consistent, reliable responses because it has been trained on the institution's actual procedures and communication norms.

Cost at volume

If 5,000 employees make an average of 3 queries per day, that is 15,000 interactions daily. At 4,000 tokens per interaction (including the RAG-retrieved context), the daily token volume reaches 60 million. Via API, that costs $5,000-$15,000 per month depending on the model. On a single GPU running an on-premise SLM, the infrastructure cost is $300-$600 per month - regardless of how many queries employees submit.

What does deployment look like?

An internal knowledge assistant is one of the fastest AI use cases to deploy because it does not require changes to existing business processes. It sits alongside the systems staff already use and adds a new, faster access path to existing information.

Document ingestion

The first step is preparing the institution's documents for retrieval. Policy documents, compliance manuals, operational procedures, and regulatory guidance are processed into chunks, converted to vector embeddings, and stored in a vector database. This creates the searchable knowledge base that the RAG layer queries.

The most important aspect of this step is document curation - ensuring the knowledge base contains current, approved versions of documents and excludes outdated or draft content. The quality of the assistant's answers is directly tied to the quality of the documents it can access.

Model configuration

Many internal knowledge assistants work well without fine-tuning at all. The RAG pipeline handles the hard part - finding the right documents - and a capable base SLM can read and summarize the retrieved content effectively out of the box.

Fine-tuning adds value when you need the model to consistently use your institution's terminology, abbreviations, and communication style. This requires preparing question-answer pairs that reflect how your staff actually ask questions and how answers should be formatted. Building this dataset takes effort - typically a few hundred curated examples gathered from real employee queries and approved answers. It is not a trivial step, but the dataset is much smaller than most people assume because you are teaching tone and style, not factual knowledge. The facts come from the RAG layer.

For a first deployment, starting with a strong RAG pipeline and an untuned base model is a practical approach. Fine-tuning can be added later once you have real usage data showing where the model's default responses fall short.

Integration

The assistant can be deployed as a standalone web interface, embedded in the institution's intranet, or integrated with existing tools like Microsoft Teams or Slack. The integration is lightweight — the assistant exposes an API endpoint that any frontend can connect to.

Ongoing maintenance

As documents are updated, the knowledge base must be refreshed. This is typically automated - new or updated documents are re-processed and re-embedded on a scheduled cadence. The SLM itself needs retraining less frequently, typically quarterly or when significant new terminology or procedures are introduced.

What results should institutions expect?

The measurable impact of an internal knowledge assistant depends on the institution's size, query volume, and baseline efficiency. But the patterns are consistent.

Faster answers

The most immediate impact is time savings. Tasks that previously required 15-45 minutes of manual search are completed in seconds. For high-volume operations teams, this translates directly into throughput improvement.

Improved consistency

When everyone queries the same system backed by the same approved documents, the answers are consistent. This reduces the variability that comes from different employees interpreting different versions of policies or applying different institutional knowledge.

Reduced load on senior staff

Experienced employees spend less time answering routine questions from colleagues. Their institutional knowledge is effectively captured in the system, making it accessible without requiring their direct involvement.

Better onboarding

New employees reach operational competence faster because they have instant access to institutional knowledge rather than relying on informal learning from colleagues. The assistant serves as an always-available, always-accurate reference.

Key takeaways

Internal knowledge management is one of the highest-impact, lowest-risk AI use cases for financial institutions. The information already exists - the challenge is making it accessible.

A purpose-built SLM connected to a RAG pipeline gives employees instant, natural-language access to the institution's policies, procedures, and regulatory guidance. It runs on the bank's own infrastructure, keeps data sovereign, and delivers consistent, sourced answers.

For institutions looking to deploy their first AI use case, or expand beyond a pilot, an internal knowledge assistant is the practical starting point. It delivers measurable value in weeks, not months, and it builds the infrastructure and organizational confidence for more complex AI deployments.

If your institution is evaluating AI for internal knowledge management, we can help you scope the deployment.

Related reading:

Ready to Own Your AI?

Stop renting generic models. Start building specialized AI that runs on your infrastructure, knows your business, and stays under your control.