KPIs and compliance metrics
1) Why compliance metrics
Metrics translate requirements and risks into manageable goals. Good KPI/KRI system:- makes the compliance status transparent and comparable over time;
- links compliance with business result (reduction of losses/fines/delays of releases);
- allows you to manage priorities and resources based on facts, not feelings;
- simplifies auditing: there are traceable formulas, sources and invariable artifacts (evidence).
- KPI - performance indicators (process efficiency).
- KRI - risk indicators (probability/impact of events).
- SLO/SLA - target service levels/term commitments.
- Leading vs Lagging: leading and lagging indicators.
2) Metrics map by domain (reference matrix)
3) Compliance North Star
1. Audit-ready in N hours (all evidence collected automatically).
2. Zero Critical Violations.
3. ≥ 90% Coverage with automated controls (policy-as-code + CCM).
4) Taxonomy of metrics
4. 1 Coverage
Control Coverage: controlled systems/all critical systems.
Evidence Coverage: artifacts collected/by audit checklist.
Policy Adoption: processes where requirements are implemented ,/all target processes.
4. 2 Effectiveness (efficacy of controls)
Pass Rate of control tests: passed/total period tests.
FPR/TPR (false/true) for detective rules.
Incidents Prevented: cases prevented by preventive controls.
4. 3 Efficiency (cost/speed)
MTTD/MTTR violations: time to detection/elimination.
Cost per Case (AML/DSAR): hours × rate + infrastructure costs.
Automation Ratio: auto-solutions/all solutions.
4. 4 Timeliness
Execution SLA (DSAR/STR/training): on time/total.
Lead Time policies: from trigger to publication.
Change Lead Time (DevSecOps gates): from PR to release for compliance checks.
4. 5 Quality (data/process quality)
Evidence Integrity:% of artifacts in WORM with hash summary.
Data Defects: errors in reg reporting/reports.
Training Score: average test score,% from the first time.
4. 6 Risk Impact
Risk Reduction Index: ∆ of total risk rate after remediation.
Regulatory Exposure: Open Critical Gaps vs License/Certification Requirements.
$ Avoided Losses (estimated): Penalties/losses averted by closing gaps.
5) Formulas and examples of calculations
5. 1 DSAR SLA
'DSAR _ SLA = (number of applications closed ≤ 30 days )/( number of applications total) '
Goal: ≥ 98%; red <95%, yellow 95-97. 9.
5. 2 Access Hygiene
'AH = obsolete _ rights (no owner/past due )/all _ rights'
Threshold: ≤ 2% (red zone> 5%).
5. 3 Drift Rate (IaC/Cloud)
'DR = drifts (IaC↔fakt mismatches )/month'
Trend: steady decline for 3 months in a row.
5. 4 Time-to-Remediate (по severity)
High: median ≤ 30 days; Critical: ≤ 7 days. Delay → auto-escalation.
5. 5 AML FPR
'FPR = false-positive _ alerts/all _ alters'
Balance with TPR and handling losses.
5. 6 Evidence Coverage (audit)
'EC = collected _ artifacts/mandatory _ by _ checklist'
Objective: 100% by the D-date of the audit; operational goal - ≥ 95% continuously.
6) Data and evidence sources (evidence)
Compliance DWH showcase: DSAR, Legal Hold, TTL, audit logs, alerts.
IAM/IGA: roles, owners, attestation campaigns.
CI/CD/DevSecOps: SAST/DAST/SCA, secret scan, licenses, gates.
Cloud/IaC: snapshots of configs, drift reports, KMS/HSM logs.
SIEM/SOAR/DLP/EDRM: correlations, playbooks, locks.
GRC: register of requirements, controls, waivers and audits.
WORM/Object Lock: unchangeable archive of artifacts + hash summary.
7) Dashboards (minimum set)
1. Compliance Heatmap - Systems × regulations × status.
2. SLA Center - DSAR/STR/training: deadlines, delinquencies, forecast.
3. Access & SoD - toxic roles, orphan accounts, progress of attestation.
4. Retention & Deletion - TTL violations, Legal Hold locks, trends.
5. Infra/Cloud Drift - IaC inconsistencies, encryption, segmentation.
6. Findings Pipeline - open/expired/closed by owners and severity.
7. Audit Readiness - evidence coverage and time to readiness "on the button."
Color zones (example):- Green - target met/stable.
- Yellow - risk of deviation, plan required.
- Red - critical deviation, immediate escalation.
8) OKR link (example quarter)
Objective: Reduce regulatory and operational risk without slowing down releases.
KR1: Increase Coverage of automated controls from 72% → 88%.
KR2: Reduce Access Hygiene from 4. 5% → ≤ 2%.
KR3: 99% DSAR on time; median response ≤ 10 days.
KR4: Drift Rate clouds − 40% QoQ.
KR5: Time-to-Audit-Ready ≤ 8 hours (dry-run).
9) RACI for metrics
10) Measurement frequency and procedures
Daily: CCM alerts, drift, secrets, critical incidents.
Weekly: SLA DSAR/STR, DevSecOps gates, Access Hygiene.
Monthly: pass rate controls, repeated findings, Evidence Coverage.
Quarterly: OKR-summary, Risk Reduction Index, audit-rehearsal (dry-run).
Threshold review procedure: trend, cost and risk analysis; changing thresholds - via Board.
11) Quality of metrics: rules
Unified semantics: dictionary of terms and SQL templates.
Formula versioning: "metric as code" (repository + review).
Reproducibility check: reperform scripts for auditors.
Immutability of artifacts: WORM + hash chains.
Privacy: minimization, masking, control of access to KPI showcases.
12) Query examples (SQL/pseudo)
12. 1 DSAR SLA (30 days):
sql
SELECT
COUNTIF(closed_at <= created_at + INTERVAL 30 DAY) / COUNT() AS dsar_sla_rate
FROM dsar_requests
WHERE created_at BETWEEN @from AND @to;
12. 2 Access Hygiene:
sql
SELECT
SUM(CASE WHEN owner IS NULL OR expires_at < CURRENT_DATE THEN 1 END)
/ COUNT() AS access_hygiene
FROM iam_entitlements
WHERE system_critical = TRUE;
12. 3 Drift (Terraform vs fact):
sql
SELECT COUNT() AS drifts
FROM drift_detections
WHERE detected_at BETWEEN @from AND @to
AND severity IN ('high','critical');
13) Thresholds (reference examples, adapt)
14) Antipatterns
Metrics "for report" without owner and action plan.
Mixing formula versions → disparate trends.
Coverage without efficiency: high Coverage, but high drift and repeated findings.
Ignores the cost of false positives (FPR) in AML/CCM.
Metrics without risk context (no association with KRI and licenses).
15) Checklists
KPI system startup
- Metrics dictionary and single "metrics as code" repository.
- Assigned owners (RACI) and refresh rates.
- Sources and the Compliance showcase are connected.
- Dashboards and color zones, SLO/SLAs and escalations are configured.
- WORM archive and report hash.
- Dry-run for audit with reperform.
Before quarterly report
- Verification of formulas, anomaly control.
- Update of near-regulatory thresholds.
- Cost/benefit analysis FPR vs TPR.
- Red Zone Improvement Plan.
16) Metrics maturity model (M0-M4)
M0 Manual accounting: Excel-tables, irregular reports.
M1 Catalogue: single showcase, basic SLAs and trends.
M2 Automated: real-time dashboards, escalation.
M3 Orchestrated: policy-as-code, CCM, auto-evidence, reperform.
M4 Continuous Assurance: "audit-ready by button," predictive (ML) risk metrics.
17) Related wiki articles
Continuous Compliance Monitoring (CCM)
Compliance and reporting automation
Risk-based audit
Policies and Procedures Lifecycle
Legal Hold and Data Freeze
DSAR: user requests for data
Data Retention and Deletion Schedules
Total
Strong compliance KPIs are clear formulas, reliable sources, owners and thresholds, an automated showcase, and deviation actions. This makes compliance a predictable service with a measurable impact on business risk and speed.