Zenoo
Risk operations

Transaction monitoring beyond sanctions: behavioural risk signals

Transaction monitoring beyond sanctions: behavioural risk signals
Zenoo's Editorial Team9 min read
Share

Most transaction monitoring programmes are designed to answer one question: is this transaction connected to a sanctioned entity? That is necessary. It is also insufficient. Sanctions screening catches known threats, people and entities that have already been identified and designated. The money launderers who are not yet on any list, the fraud patterns that do not match any predefined rule, the structuring behaviours that stay just below your alert thresholds: these require a different approach.

Behavioural transaction monitoring analyses how customers transact over time and identifies patterns that deviate from expected behaviour. It is the difference between checking a name against a list and understanding whether a customer's financial activity makes sense given who they claim to be.

Why sanctions screening alone is not enough

Sanctions screening is a binary check at a single point in time. Is this name on a list? Yes or no. It is essential for compliance, but it addresses only one dimension of money laundering risk.

Consider a customer who passes all sanctions, PEP, and adverse media screening at onboarding. Their identity is verified. Their risk assessment is completed. They are categorised as standard risk. Six months later, their transaction behaviour changes fundamentally. Transaction volumes double. New counterparties appear in jurisdictions they have never transacted with before. The size distribution of their transactions shifts, clustering just below reporting thresholds. None of this triggers a sanctions alert because no sanctioned entity is involved. But the behaviour pattern is consistent with money laundering typologies that FATF and national FIUs have identified as high-risk.

A sanctions-only monitoring programme misses this entirely. A behavioural monitoring programme catches it.

"We had a customer who passed every screen at onboarding and every annual re-screening for three years. Clean as a whistle. But when we implemented behavioural monitoring, we flagged them within two weeks. Their transaction patterns had shifted six months earlier, and nobody noticed because our monitoring was entirely list-based. The behavioural signals were obvious once we were looking for them."

The building blocks of behavioural monitoring

Effective behavioural transaction monitoring rests on four pillars.

Customer baselines. Before you can identify unusual behaviour, you need to define what usual behaviour looks like. A customer baseline is a profile of expected transaction activity derived from the customer's declared business purpose, risk category, and historical transaction patterns. A freelance consultant who typically receives two or three payments per month from UK clients has a very different baseline from an import/export business transacting daily across multiple jurisdictions.

Building accurate baselines requires sufficient transaction history (typically 3 to 6 months of activity) and categorisation that reflects the customer's genuine business model. Generic baselines (e.g., "all retail customers share the same expected pattern") are too coarse to detect meaningful deviations. The more granular your baselines, the more useful your monitoring.

Deviation detection. Once baselines are established, the monitoring system needs to identify when a customer's behaviour deviates from their baseline. The key metrics include: transaction volume (number and total value), counterparty patterns (new counterparties, jurisdictions, concentration), transaction size distribution (changes in average size, clustering near thresholds), velocity changes (acceleration or deceleration of activity), and channel usage (shifts between payment methods or platforms).

Each deviation is not necessarily suspicious on its own. A customer's transaction volume might increase because their business is growing. The monitoring system needs to assess deviations in context, combining multiple signals to distinguish genuine changes in business activity from patterns that warrant investigation.

Typology matching. FATF, national FIUs, and industry bodies publish money laundering and terrorist financing typologies: documented patterns of behaviour associated with specific types of financial crime. Your monitoring system should be configured to detect these typologies, including structuring (splitting transactions to avoid reporting thresholds), layering (moving money through multiple accounts or entities to obscure its origin), round-tripping (sending money abroad and receiving it back through a different channel), and rapid movement (funds entering and leaving an account within a very short period with no economic rationale).

Network analysis. Individual transaction analysis looks at each customer in isolation. Network analysis looks at connections between customers. If two of your customers are transacting with each other in patterns that suggest a coordinated structure, or if a group of your customers share counterparties in a pattern consistent with a laundering network, individual monitoring will not catch it. Network analysis maps these relationships and identifies suspicious patterns across your customer base.

Building custom risk scoring algorithms inside an orchestration layer

One question that comes up repeatedly with teams moving beyond sanctions-only monitoring: can we define our own risk model, or are we stuck with whatever logic the vendor ships?

The honest answer in most legacy platforms is the latter. Risk scoring is hard-coded by the vendor, the weighting is opaque, and "customisation" means picking from a small menu of preset profiles. For an MLRO who needs to weight transaction monitoring signals, screening outcomes, and KYC attributes in a way that reflects their specific book of business, the menu approach falls short.

An orchestration platform with a configurable rules engine handles this differently. The compliance team defines the scoring logic ad hoc: combine sanctions and PEP outcomes, adverse media hits, KYC and KYB attributes, behavioural deviation scores, and transaction-level signals into a weighted composite. Each input has its own weight. Conditional logic handles segment-specific rules ("score this differently for corporate customers in jurisdiction X"). Score thresholds drive automated decisions, escalations, or step-ups.

The crucial bit is governance. A custom scoring model is only useful if it is testable, auditable, and reversible:

  • Shadow mode runs the new model alongside the live model for a defined period. The team sees what would have happened without changing real outcomes.
  • Versioning captures every change to the model with a reason and an approver. Six months later, if a regulator asks why a specific customer scored where they did, the version that was live at the time is reproducible.
  • Champion or challenger testing runs an updated model against the existing one on live traffic, with clear metrics for promotion or rollback.

This is what "build your own risk model" should mean in practice. Not a config wizard with three sliders. The ability to express the model your business actually needs, prove it works, and change it without filing an engineering ticket.

Calibration: the art and science of thresholds

The biggest operational challenge in behavioural monitoring is calibration. Set your thresholds too tight, and every minor fluctuation generates an alert. Set them too loose, and genuine suspicious behaviour goes undetected. Getting the balance right is an ongoing process, not a one-time configuration.

The calibration methodology should start with typology-based thresholds: configure specific detection rules for known money laundering patterns based on FATF and FIU typologies. These rules should be tight because they target specific, documented risks.

Then add statistical deviation thresholds: flag customers whose behaviour deviates from their baseline by more than a defined number of standard deviations. These thresholds should be looser than typology-based rules because they are catching unknown patterns, and the false positive rate will be higher.

Finally, implement contextual filters that suppress alerts when the deviation has an obvious explanation. If a customer's transaction volume increases at the same time as their industry's seasonal peak, that is probably not suspicious. If it increases at a time when their industry is typically quiet, it might be.

"Our first attempt at behavioural monitoring generated 8,000 alerts in the first month. Our team could handle about 500. So we spent three months calibrating: tightening thresholds on some detection rules, loosening others, adding contextual filters, and improving our baselines. By month four, we were down to 600 alerts per month with a genuine hit rate of about 12%. That was useful. 8,000 alerts was not."

Integrating behavioural signals with risk ratings

Behavioural monitoring should not exist in isolation. The signals it generates should feed into your customer risk assessment framework, influencing risk ratings in real time rather than waiting for the next scheduled review.

When a customer's behaviour triggers a monitoring alert, several things should happen. First, the alert is investigated and dispositioned by an analyst (or an AI agent, for routine cases). Second, if the investigation confirms that the behaviour is genuinely unusual, the customer's risk rating should be recalculated to reflect the new information. Third, if the recalculated risk rating triggers a higher risk tier, the customer should be subject to the enhanced due diligence and monitoring associated with that tier.

This creates a feedback loop: monitoring detects behavioural changes, risk ratings update to reflect those changes, and monitoring intensity adjusts to the new risk level. A customer who starts behaving in ways inconsistent with their profile receives more scrutiny. A customer who has been flagged but subsequently returns to normal patterns can have their monitoring intensity reduced.

graph TD
    A["Customer behaviour changes"] --> B["Alert generated by monitoring system"]
    B --> C["Analyst investigation"]
    C --> D{Behaviour confirmed unusual?}
    D -->|Yes| E["Risk rating recalculated"]
    D -->|No| F["Alert dispositioned as false positive"]
    E --> G{New risk tier triggered?}
    G -->|Yes| H["Apply enhanced due diligence"]
    G -->|No| I["Continue standard monitoring"]
    H --> J["Increased monitoring intensity"]
    I --> K["Normal monitoring intensity"]
    F --> K
    J --> L["Customer returns to normal behaviour"]
    L --> M["Risk rating reassessed, intensity reduced"]
    M --> K

Integrating alerts from your existing transaction monitoring system

Behavioural monitoring does not have to mean ripping out what you already have. Plenty of firms run a perfectly serviceable in-house transaction monitoring system, or a third-party tool whose alert logic the team has invested years in tuning. The question is not whether to replace it. The question is whether those alerts can flow into a unified customer view alongside KYC, KYB, screening, and case management.

The integration pattern is straightforward when the orchestration layer is designed for it. Inbound alerts arrive over REST API or webhook with a defined payload:

  • Customer identifier (so the alert attaches to the right profile)
  • Alert type and code (e.g., "structuring threshold breach", "velocity anomaly", "counterparty concentration")
  • Severity
  • Timestamp
  • Narrative or supporting context from the originating system
  • Optional: source transaction IDs and supporting evidence links

Once the alert lands, it does the work that an alert from any other source would do. It attaches to the customer's profile in the unified case management view. It triggers a workflow (assign analyst, set SLA, queue for review). It can re-fire risk scoring, push the customer up a tier, or escalate to enhanced due diligence depending on the rules the team has configured. The investigator sees the alert in the same place they see screening hits, KYB changes, and onboarding flags.

What changes for the team running the in-house system: nothing about the detection logic. They keep their tuning, their rules, their proprietary signals. What they get is a unified case management surface so investigators do not bounce between four browser tabs to assemble the picture of a single customer.

This is also a useful migration pattern. Teams who eventually want to fold transaction monitoring into the same orchestration layer can do it gradually. Start by piping alerts in. Layer on behavioural monitoring built inside the orchestration layer. Run them in parallel. Decommission the legacy system when the new behavioural model has earned the team's confidence. No big-bang cutover, no period of monitoring blindness.

Technology requirements

Behavioural transaction monitoring at scale requires specific technology capabilities that many compliance teams lack.

Capability Why it matters Implementation challenge
Real-time data processing Transaction data must be processed in real time or near real time. Batch processing with overnight runs means monitoring yesterday's behaviour. For time-sensitive typologies like rapid movement of funds, overnight processing is too slow. Requires streaming architecture and continuous compute infrastructure
Machine learning capabilities Rule-based detection catches known typologies, but machine learning models identify patterns no predefined rule would catch. Unsupervised learning can detect anomalous behaviour without being told what to look for, making it valuable for emerging techniques. Requires data science expertise and labelled training data for model development
Flexible storage and compute Behavioural baselines require storing and analysing months or years of transaction history for every customer. For firms with large customer bases and high transaction volumes, this is a significant data engineering challenge. Scales with customer base size and transaction throughput. Storage costs increase with data retention requirements.
Integration with compliance stack Monitoring alerts need to flow into case management, trigger risk recalculations, and feed into SAR filing workflows. Disconnected systems mean managing behavioural monitoring manually, which overwhelms any benefit. Requires API connectivity and data mapping across multiple vendor platforms

Getting started

If your current monitoring programme is primarily sanctions-based and you want to add behavioural capability, here is a practical starting point.

Start with your highest-risk customers. Build behavioural baselines for your high-risk customer segment first. This is where the monitoring will have the greatest impact, and the smaller population makes calibration more manageable.

Focus on three typologies. Do not try to detect everything at once. Pick three money laundering typologies that are most relevant to your business model and configure detection rules for those. Structuring, rapid movement, and counterparty concentration are good starting points for most financial institutions.

Accept a learning period. Your first few months of behavioural monitoring will require intensive calibration. False positive rates will be high initially. Baselines will need refinement. Detection rules will need tuning. This is normal. Budget the analyst time for the calibration period and do not judge the programme's effectiveness until calibration is complete.

Key takeaways

  • Sanctions screening alone catches only known threats already on watchlists. Behavioural monitoring detects suspicious patterns that do not match predefined rules, including structuring, layering, and rapid movement of funds.
  • Effective behavioural monitoring requires four pillars: granular customer baselines, deviation detection across multiple metrics, typology matching against FATF and FIU patterns, and network analysis to spot coordinated suspicious activity across your customer base.
  • Calibration is an ongoing process, not a one-time setup. Typology-based thresholds should be tight, statistical deviation thresholds looser, and contextual filters should suppress false positives tied to legitimate business events.
  • Behavioural signals must integrate into your risk rating framework in real time, triggering recalculations and enhanced due diligence automatically rather than waiting for scheduled reviews.
  • Begin with high-risk customers only, focus on three relevant typologies, and expect a 3 to 4 month calibration period where false positive rates are high. Judge success only after baselines and rules are refined, not in month one.

Sanctions screening is necessary but insufficient. The threats that sanctions lists do not capture, the patterns that predefined rules cannot anticipate, and the behavioural changes that emerge over time all require a monitoring approach that looks beyond names and lists. Behavioural transaction monitoring is that approach. It is more complex to implement and calibrate than list-based screening, but it catches risks that screening alone never will.

If you want to understand how behavioural monitoring fits into an orchestrated compliance platform, talk to us. 30 minutes. Your data. No slides.

Was this useful?
Share
Z

Published by

Zenoo's Editorial Team

Practical, unbiased content on KYC, AML, and compliance operations. Written by the team building tools to make compliance work better.

The compliance intelligence you actually need

Weekly insights on KYC, AML, and compliance operations. No vendor spin. No gated whitepapers. Just honest, useful guidance.

More from Zenoo Insights

22 hours per alert is too long. Cut it to 12 minutes.

One platform. 10 AI agents. 240+ check types. Live in weeks, not months.

30 minutes. Your data. No slides.