Single Blog

  • Home
  • TMS Review Checklist That Holds Up in Audit
TMS Review Checklist That Holds Up in Audit

TMS Review Checklist That Holds Up in Audit

March 4, 2026

A regulator rarely criticises a firm for having a transaction monitoring system (TMS). They criticise it for what it fails to do in practice: miss obvious typologies, generate unmanageable alert volumes, or produce no defensible rationale for why key risks were accepted.

A transaction monitoring review, done properly, is not a software inspection. It is an operational and governance test of whether your monitoring is aligned to your Business Risk Assessment, customer risk model, product design, and actual transaction behaviour. The outcome you want is simple: monitoring that is risk-based, explainable, and demonstrably effective.

How to use a transaction monitoring system review checklist

This transaction monitoring system review checklist works best when you treat it as a set of control questions you can evidence, not a shopping list of features. For each area, ask two things: (1) what is our documented position, and (2) what can we prove with data, tickets, logs, minutes, testing results, and MI.

If your firm is small, you may not have separate teams for typology design, tuning, and QA. That is fine. The trade-off is independence. Where you cannot separate duties, you compensate with stronger documentation, sampling, and sign-off.

1) Governance and accountability: who owns what, and can you show it?

Start with ownership because most weaknesses cascade from here. Regulators expect clear allocation of responsibility across first and second line, and an MLRO function that can challenge, not just receive statistics.

You should be able to point to a named owner for scenario design and changes, a named owner for operational alert handling, and a clear approver for material tuning decisions. If your vendor “manages” tuning, you still own the risk. The practical question is whether you can evidence oversight: meeting minutes, change requests, approvals, and a risk acceptance process where you have chosen not to implement a detection because of cost, feasibility, or disproportionate impact.

Also test whether your governance triggers make sense. A quarterly review cadence can be appropriate for stable products with low change, but it is weak for fast-growing fintech models, new corridors, new payment rails, or periods of known typology shifts.

2) Risk alignment: does monitoring reflect your BRA and customer risk model?

A common audit finding is misalignment between documented risks and what the TMS actually monitors. Take your Business Risk Assessment and map each material inherent risk to detection coverage. Do not aim for perfection. Aim for a defensible story.

If your BRA highlights high exposure to remote onboarding, cross-border flows, cash-like instruments, or third-party payments, you should be able to show which scenarios, thresholds, and investigative playbooks address those risks. Where monitoring cannot reasonably detect a risk, you should show the compensating controls (for example, enhanced CDD triggers, sanctions screening, or stricter source of funds requirements).

Be careful with inherited templates. A generic “structuring” rule may be irrelevant if your product does not accept cash, while you may have a gap on rapid account funding and instant withdrawal behaviour that is far more relevant.

3) Data integrity and coverage: the system is only as credible as the inputs

Most TMS failures are data failures dressed up as monitoring failures. Review what data the system receives, at what frequency, and with what quality controls.

Focus on completeness first: are all relevant transaction types captured (including reversals, chargebacks, refunds, internal transfers, crypto on-ramps, wallet-to-wallet movements, and off-platform settlement if applicable)? Then assess enrichment: do you have reliable customer identifiers, counterparty details, country codes, channel indicators, merchant category codes, device information, and relationship mappings?

Next, test for silent breaks. File drops that fail without alerting anyone, mapping changes that shift values into the wrong fields, time zone issues that distort “velocity” scenarios, and duplicate records that inflate volumes. You want reconciliations between core ledgers and monitoring feeds, plus controls that notify owners when a feed is late or malformed.

If you cannot evidence data lineage and control points, your monitoring decisions become hard to defend, even if investigators are doing good work.

4) Scenario and typology coverage: can you articulate what you detect and why?

A strong review does not start with the number of rules. It starts with typologies relevant to your business model and delivery channels.

Document your core typologies and link them to scenarios. For example: layering through multiple counterparties, mule activity patterns, rapid movement in and out, pass-through behaviour, unusual use of high-risk corridors, abuse of refunds, or transaction behaviour inconsistent with stated profile. Each should have a rationale, key risk indicators, and a clear description of what the scenario is intended to catch.

Check for overlap and contradiction. Too many similar scenarios inflate alerts without improving detection. Conversely, overly broad scenarios can create “noise” that desensitises investigators. The right balance depends on your risk appetite and resources, but you should be able to justify it.

Also confirm whether you have scenarios tailored for higher-risk segments, such as PEPs, high-risk third countries, cash-intensive businesses, or complex corporate structures. One-size-fits-all monitoring is rarely risk-based.

5) Thresholds, calibration, and tuning: decisions you can defend

Tuning is where good intentions meet operational reality. Review how thresholds were originally set, how they are adjusted, and who approves changes.

A defensible tuning approach includes historical analysis, segmented thresholds (not just one global number), and documented trade-offs between false positives and missed detection. If your alert rate is extremely low, that is not automatically good. It may indicate thresholds set so high that you are not detecting anything meaningful. If your alert rate is extremely high, investigators may be closing alerts too quickly to keep up.

Look for evidence of periodic calibration: back-testing against known suspicious cases, testing on samples of historical transactions, and review of “near misses” where suspicious activity was identified through other means (for example, adverse media, law enforcement requests, or relationship manager escalation) but the TMS did not alert.

Where you accept limitations – for example, you cannot monitor certain off-platform movements – record the rationale and compensating controls. Risk acceptance is not a weakness when it is explicit, approved, and monitored.

6) Alert workflow and investigations: operational control, not just case closure

Alert handling is often the most visible part of a regulator’s file review. You need a workflow that produces consistent, complete investigative outcomes.

Test whether alerts are triaged using clear criteria, and whether escalations are timely. Review the quality of narratives: do investigators explain what they reviewed, what they found, and why they concluded as they did? “No adverse found” is not an investigation. A good narrative references specific account behaviour, counterparties, KYC/CDD context, and supporting evidence.

Quality assurance matters. If you do not have a dedicated QA team, implement a second-review sample for high-risk alerts and high-risk customers, and keep records of feedback loops. Regulators will look for evidence that you identify root causes of poor investigations and fix them through training or process changes.

Also review case ageing. Long-open alerts can signal resourcing gaps or unclear decisioning. A small backlog may be acceptable; a sustained backlog in high-risk segments is harder to justify.

7) SAR/STR decisioning and record-keeping: consistent thresholds and rationale

Your monitoring output should feed a consistent reporting process. Review the criteria for escalating cases to the MLRO and the thresholds for filing. In Malta, this typically means STR obligations with the FIAU, but the principle applies across jurisdictions.

Test whether reporting decisions are documented with the same discipline as investigative decisions. The file should show why the activity is suspicious, what alternative explanations were considered, what additional information was sought, and who approved the final decision.

Keep an eye on defensive non-reporting. Some firms close too many cases as “plausible” without evidencing why. Others over-report to be safe, which can indicate weak investigation quality. The right position depends on your risk exposure, but consistency and rationale are non-negotiable.

8) Management information: can leadership see effectiveness, not just volume?

MI should help you manage risk, not just count alerts. Basic volumes by scenario and by team are a start, but they do not show effectiveness.

Include measures such as: alert-to-case conversion, case-to-escalation rates, outcomes by risk segment, typologies detected, repeat subjects, time-to-triage, time-to-close, and QA findings. If you have a customer risk model, overlay MI by risk rating to confirm you are not spending most of your effort on low-risk populations.

MI should also capture change events – new products, new corridors, policy updates – so you can explain shifts in alert behaviour. A spike in alerts is only meaningful if you can connect it to a cause and a response.

9) Change management and model risk controls: controlled change, tested change

Scenario changes, threshold adjustments, and data mapping updates should be controlled like any other material change. Even if your organisation does not call it “model risk management”, the same expectations apply: you should not change detection logic without testing and approval.

Maintain a change log that includes what changed, why, what testing was done, what results were observed, and who signed off. For significant changes, keep evidence of pre- and post-change performance. If you rely on a vendor for updates, ensure you receive release notes and can evidence your review of the impact.

Also confirm access controls. Who can edit scenarios? Who can change risk parameters? Admin access without oversight is a recurrent control weakness.

10) Audit defensibility: can you recreate decisions and prove ongoing review?

When an auditor or regulator asks “why did you do this?”, you need to answer with evidence, not memory.

Ensure you can recreate an alert from the underlying transaction data, show what the system flagged, show what the investigator reviewed, and show how the decision was reached. Retention periods should align to applicable AML record-keeping requirements, and your retrieval process should be tested. If it takes weeks to find a case file or you cannot retrieve supporting documents, you have an operational risk as well as a compliance risk.

Finally, test whether your monitoring programme is reviewed end-to-end, not just tuned. A yearly independent review can be proportionate even for smaller firms, particularly where growth has been rapid or where the business model carries higher inherent risk. If you need an independent perspective, Complipal supports monitoring programme reviews that focus on practical effectiveness and clear remediation actions, rather than checkbox commentary.

A helpful way to think about transaction monitoring is this: the goal is not to build the most complex system, but to build one you can explain, operate, and improve with discipline as your risk profile changes.