IE 11 Not Supported

For optimal browsing, we recommend Chrome, Firefox or Safari browsers.

Using Secure Small Language Models to Navigate Big Data Sets

How agencies can use on-premises AI models to detect fraud faster, prove control effectiveness and turn overwhelming data volumes into clear investigative leads — all through a simple chat interface.

Hands on an open laptop with a digital image on top of data being extracted from a database.
Adobe Stock/NicoElNino
Every day, government agencies generate millions of audit log entries tracking system access, financial transactions and data changes across dozens of critical systems. Internal auditors and investigators face an impossible task: finding fraud signals buried in massive data sets using manual reviews and basic spreadsheet tools. By the time suspicious patterns surface, the damage is often done.

Small language models (SLMs) running on agency-owned servers are changing this reality. These AI systems let auditors ask direct questions in plain language and receive instant analysis of complex patterns that would take weeks to uncover manually. The technology is secure, practical and available now — without sending sensitive data to external cloud services.


THE BREAKING POINT FOR MANUAL LOG REVIEW


Traditional audit approaches cannot keep pace with modern government operations. A single financial system might log 100,000 transactions daily, while access logs from a citizen database could generate another 500,000 entries. Auditors using spreadsheets and SQL queries spend 60 to 80 percent of their time just extracting and organizing data before analysis can even begin.

The real cost shows up in delayed detection. When an auditor finally spots duplicate payments to the same vendor three months after they occurred, the fraud scheme has often expanded. Manual reviews catch obvious anomalies — like exact duplicate transactions or access at 3 a.m. — but miss sophisticated multistep patterns where each individual action appears normal.

This gap between audit capacity and system activity creates genuine risk for agencies managing sensitive citizen information and large financial flows.

HOW SLMs TRANSFORM AUDIT INVESTIGATION


Unlike large language models requiring massive cloud infrastructure, SLMs are compact AI systems with 7 to 14 billion parameters that run efficiently on a single GPU server. Models like Meta’s Llama 3.1 or Mistral 7B provide the reasoning capability needed for fraud detection while keeping all data processing inside agency walls.

The core breakthrough is conversational analysis. Instead of writing complex SQL queries, an auditor types, “Show me any users who accessed confidential records after hours and also made unusually large payments within 48 hours.” The SLM interprets this natural language request, generates the appropriate database queries, analyzes the results and presents findings with supporting evidence — all in under 30 seconds.

This automation eliminates the technical barrier that currently limits audit effectiveness. Auditors can focus on investigating findings and asking follow-up questions rather than wrestling with data extraction. 

SLMs excel at recognizing complex behavioral sequences that escape human notice because individual steps appear benign. Consider this pattern: An employee logs in at an unusual time, accesses a sensitive client record they’ve never viewed before, then executes a rarely used transaction type for a round dollar amount. Each action alone might not trigger an alert, but the sequence is a clear fraud indicator.

Traditional rule-based systems would need explicit programming for every possible suspicious combination. SLMs recognize these patterns through their training on vast data sets describing fraud schemes and anomaly detection principles. The models connect separate events into unified behavioral signals, providing early and unambiguous detection.

This capability proves especially valuable for identifying insider threats and sophisticated fraud schemes designed to evade conventional controls.

THE ARCHITECTURE THAT MAKES THIS PRACTICAL


Government agencies rightfully approach AI with caution given the sensitivity of audit data. The deployment model addresses these concerns through three core principles.

First, the SLM runs entirely on agency-controlled hardware — either in the data center or an approved government cloud environment. No audit data ever transmits to external AI services or training systems. The model only analyzes information the agency explicitly provides from its own secure databases.

Second, a control layer sits between the auditor and the AI model, enforcing all access policies and compliance requirements. This framework determines exactly which logs the model can see, what questions are permissible and which data elements must remain masked. It’s rule-based governance, not AI decision-making.

Third, the system operates under strict human-in-the-loop principles. All AI findings are presented with supporting evidence for verification by trained auditors before any action is taken. The technology accelerates investigation; it doesn’t replace professional judgment.

Agencies can deploy this capability without replacing existing systems or undertaking multiyear modernization projects. The framework connects to current audit log sources through standard database interfaces and encrypted APIs.

A typical setup requires one GPU server costing $8,000 to $15,000 with sufficient processing power to support five to 10 concurrent auditors analyzing millions of log entries. The SLM software itself is open source and available without licensing fees.

The implementation path follows a practical three-phase approach. Start with a single high-value audit log table and two to three auditors testing the system for one to two months. This pilot validates accuracy and builds confidence before expanding to additional data sources. Full production deployment typically occurs within six months, followed by optional fine-tuning of the model on agency-specific fraud patterns.

Early adopters report that pretrained SLMs achieve 70 to 80 percent accuracy on fraud detection tasks immediately, without any customization. With fine-tuning on agency data over six to 12 months, accuracy improves to 85 to 95 percent for common fraud types.

The time savings are substantial. Tasks that previously required eight hours of query writing and data manipulation are now complete in minutes. This frees investigators to handle larger case volumes and conduct more thorough follow-up on flagged issues.

However, AI-powered auditing is not foolproof. Even with structured database queries, SLMs occasionally misinterpret complex scenarios or miss subtle patterns. This is why human verification remains essential: The technology augments professional expertise rather than replacing it.

THE PATH FORWARD FOR MODERN AUDITING


AI-powered audit log review strengthens security posture and improves detection speed without requiring agencies to trust external AI services with sensitive data. The technology is mature enough for production use, affordable enough for most budgets and straightforward enough to implement within existing operations.

For agencies seeking to modernize internal audit capabilities while maintaining full data control, SLMs offer a practical and secure solution. The question is no longer whether AI can improve audit effectiveness — it’s how quickly agencies will adopt the tools already available.

Raj Prahast Maheswarapu is a principle architect for data safeguarding and governance at the Florida Department of Revenue, and is a public-sector technology leader specializing in data safeguarding, fraud prevention and AI-driven modernization for government programs. He has presented on AI governance and fraud detection at the PMI Global Summit, Georgia Digital Government Summit and Florida Digital Government Summit (Digital Government Summits are hosted by Government Technology). His work focuses on strengthening data integrity and protecting sensitive citizen information in high-volume government environments.