Dmitry Reznik

Written by: Chief Technology Officer

Dmitry Reznik

Posted: 02.04.2026

16 min read

Once you’ve got there, here is a quick primer on what likely interests you.

The Digital Operational Resilience Act is a core North Star for the financial entities in the European Union. For over a year, this regulation has been controlling the adherence to the technical and operational standards in the industry.

In 2026, most fintech teams have two relatively new challenges:

1. Documents → real actions (read, software tests): It all begins with policies (almost all). But new laws demand something more — they want you to ensure test logs, the results of regular disaster recovery exercises, and failover metrics. Without this evidence, passing a regulatory audit becomes extremely difficult. Penalties vary depending on the severity of non-compliance.

2. Independent third-party vendors → comprehensive partnerships: DORA law makes financial companies responsible for the uptime and security of their vendors. Your external vendor made a mistake in their security protocols? Your business can get cascading system failures and regulatory fines due to your partner’s misstep.

Diagram showing key compliance implications of DORA, including ICT risk management, incident reporting, digital operational resilience testing, third-party risk management, cross-border considerations, enforcement and penalties, and governance implications.
Key compliance implications of DORA for financial institutions

Moreover, fragmented execution is visible since supervisory authorities, ESAs, and national regulators are coordinating oversight.

In this blog, we explain what the DORA regulation is and turn its outputs into actionable engineering and quality outtakes.

Let’s start with the most logical step — the DORA regulation summary.

What the DORA regulation is (and what it is not)

Before the DORA (Digital Operational Resilience Act), European financial entities had to figure out ICT risk expectations across different member states. Regulatory demands hugely varied by country, and cross-border operations were difficult to manage (to audit as well, though).

That regulation 2022/2554 covers five connected pillars:

1.

ICT risk management

2.

ICT-related incident reporting

3.

Digital operational resilience testing

4.

ICT third-party risk management

5.

Information-sharing arrangements

Unfortunately, many sources still describe the Act as a cybersecurity regulation. Not precise and may lead to failed audits. To be more precise, the Act demands:

  • Governance and control: Leadership boards are accountable for ICT risk management and execution.

  • Business continuity planning: You need tested failover strategies backed by data.

  • Resilience testing: Threat-led penetration testing (TLPT) and automated operational drills are now mandatory.

  • Third-party vendor oversight: As a financial project, you remain fully accountable for risks introduced by your ICT third-party providers.

Who must comply

The full scope is detailed in Article 2 of the Regulation. Critical ICT third-party service providers (CTPPs), cloud infrastructure providers, data analytics platforms, and SaaS vendors now face direct supervision from authorities (ESMA).

Here are the main recipients of DORA compliance:

  • credit institutions

  • payment institutions and electronic money institutions

  • investment firms

  • insurance and reinsurance undertakings

  • crypto-asset service providers (where applicable under EU financial regulation)

  • central counterparties, trading venues, and other market infrastructure entities

If your company builds software or provides IT services to a regulated European financial firm, you operate inside their supply chain. Chances are, your upcoming vendor contracts will include DORA requirements, annual audit cycles, and procurement requests.

If you don’t provide test execution logs, incident response metrics, or disaster recovery results, you are not going to get new deals, yet you may get the termination of existing partnerships.

The 5 pillars of DORA

As per PwC, 22,000+ financial establishments and IT service providers must comply with the regulation. Non-compliance brings about fines of up to EUR 5M or 3-12.5% of annual turnover for serious breaches. For comparison, the GDPR fine baseline is up to 4% or EUR 20M of annual turnover.

Circular diagram illustrating the five DORA pillars: ICT risk management, incident management, digital operational resilience testing, ICT third-party risk management, and information sharing, with detailed subcomponents for each pillar.
Overview of DORA pillars and key operational areas

Any figures make compliance efforts feasible.

Pillar 1. ICT risk management

What DORA expects

Your go-to guidance here is Articles 5-16. In a nutshell, they cover governance for technology risks: maintain a precise asset inventory, change and release controls, backup and restore capabilities.

What teams must implement

  • A complete inventory of all internal and external information assets

  • Change management protocols to block unauthorized code deployments

  • Scheduled and automated backups

  • Continuous monitoring tools to detect system anomalies

What to test

1.

Run a data recovery plan to check backup integrity and restoration speed

2.

Trigger simulated system spikes to validate monitoring alerts

3.

Review release pipelines to confirm that change controls function correctly

Evidence artifacts

  • An updated risk register

  • A complete asset inventory map

  • Access review logs

  • Restore test execution results

Pillar 2. ICT-related incident management and reporting

What DORA expects

The new approach defines how tech teams must classify and report ICT-related incidents: timelines, report content, and templates. Also, the reporting becomes direct to competent authorities, according to the technical standards adopted by the European Supervisory Authorities.

What teams must implement

  • Automated incident detection and response mechanisms

  • The workflow should include the following stages: detect → triage → classify → internal comms → regulatory reports → post-incident review

What to test

1.

How fast can you detect and classify a simulated critical failure

2.

If your automated report generation meets the required ITS templates

3.

Whether your senior management gets alerts properly and immediately in case of an issue

Evidence artifacts

  • Incident classification logs

  • Completed regulatory reporting templates (using mock drill data or historical incidents)

  • Post-incident review documentation detailing root causes

Pillar 3. Digital operational resilience testing

What DORA expects

A testing structure based on a risk, criticality, and size profile: routine controls testing, scenario-based testing, and advanced Threat-Led Penetration Testing (TLPT). You can dive deeper into mandatory testing programmes in Articles 24-27.

What teams must implement

  • Automate routine QA and security regression tests in the CI/CD pipeline

  • Schedule scenario-based disaster recovery actions

  • Advanced checks; for example, TLPT for designated entities (you can use certified external red teams for that)

What to test

Measure how quickly systems recover and how well defenses hold up against mock attacks. Track every defect discovered during these tests and measure remediation speed.

Evidence artifacts

  • Logs detailing specific defects found during resilience tests

  • Remediation tracking boards showing ticket closures

  • Re-test proof confirming the identified vulnerability is securely closed

Pillar 4. Third-party ICT risk management

What DORA expects

Get on the same page with your vendors. In any sense: in cross-checking critical functions, enforcing specific contract clauses, in ongoing performance monitoring, and in crisis strategies. Learn more about vendor oversight in Articles 28-44.

What teams must implement

  • Determine the most critical business functions, tie them to your software, and grade them

  • Categorize the entire software supply chain based on that graduation

  • Include mandatory audit rights in all vendor contracts. Re-sign the existing ones if needed

  • Ensure exit strategies to migrate away from a failing vendor

What to test

1.

Audit vendor incident response times through joint trainings

2.

Simulate a critical vendor outage to test the technical feasibility of your exit plan

3.

Validate vendor participation in your internal resilience tests

Evidence artifacts

  • SOC2 reports or equivalent vendor security certifications

  • Signed SLAs and incident notification agreements

  • Executed audit rights clauses within active vendor contracts

  • Proof of vendor participation in operational resilience testing

Pillar 5. Information sharing

What DORA expects

Your contribution to the healthy fintech industry by sharing cyber threat information and intelligence among other financial companies and institutions. When this info is standardized, it helps to ensure collective resilience across the sector.

What teams must implement

  • Secure channels to receive and share anonymized threat intelligence

  • External threat data feeds should be directly in internal security monitoring tools

What to test

1.

Whether ingested threat feeds correctly trigger internal security alerts

2.

Internal procedure for stripping personal data and anonymizing threat metrics

Evidence artifacts

  • Information-sharing policy

  • Membership documentation

  • Threat intelligence intake records

  • Action logs showing how shared intelligence improved controls

DORA “Level 1 vs Level 2”

Compliance with the regulation text isn’t enough (...if you want to complete the audit for sure). The Level 1 is a so-called DORA law; it contains core principles and obligations:

  • ICT risk management requirements

  • Incident reporting duties

  • Digital operational resilience testing

  • Third-party ICT risk management

  • Information-sharing conditions

Basically, it’s a summary of what must exist and who is accountable for that. But it’s not all what the DORA regulation is.

Level 2 discloses more — the Regulatory Technical Standards (RTS) and Implementing Technical Standards (ITS). These documents explain exactly how to execute the rules and what evidence auditors will request:

  • Incident classification thresholds and reporting templates

  • Timing requirements for initial, intermediate, and final reports

  • Criteria for threat-led penetration testing (TLPT)

  • Third-party oversight expectations

Level 1 tells you what, and Level 2 details how exactly. Let’s now break down the preparation roadmap.

An implementation roadmap for the DORA European regulation

The entire preparation may take about 6 months to set up and continuous operational effort (or until the regulation loses its power). You can target time estimations below, but always remember your capacity + some buffer on the off chance.

Phase 1:Baseline and gap assessment (2 to 4 weeks)

Objective: To understand your current operational capacity and evidence maturity and compare them with regulatory standards.

What to do:

1.

Identify critical functions

2.

Internally audit systems, data flows, and dependencies for those functions

3.

List ICT third-party providers linked to them (we’d suggest an internal database to ease tracking and management)

4.

Review that across the 5 pillars

What you get as a result:

  • Critical services register

  • System and vendor inventory with ownership

  • Control maturity assessment

  • Evidence gap log

The common situation is when teams have controls, but they actually tested them once, at the very beginning.

Phase 2: Remediation plan and governance (1 to 3 months)

Objective: To understand who will be accountable for each next phase and who will lead your engineering backlog.

What to do:

1.

Choose executive and operational owners per pillar

2.

Align the incident classification matrix to RTS edges

3.

Design a structured incident reporting playbook

4.

Define an annual resilience testing program

5.

Meet vendor contracts with DORA requirements

6.

Ensure you have a board-level reporting format for ICT risk

What you get as a result:

  • Approved remediation roadmap

  • Incident reporting operating model

  • Testing calendar

  • Vendor classification matrix

  • Updated governance charter

Clarity at this stage allows 90% of the control later.

Phase 3: Make it a routine (3 to 6 months)

Objective: To implement all previous preparation in the day-to-day routine: code, environments, testing/reaction training, performance validation, etc.

What to do:

1.

Start/improve change and access controls

2.

Run backup restoration tests against defined RTO/RPO targets

3.

Conduct resilience exercises. Target different crisis scenarios, not just the most possible ones

4.

Conduct live operational resilience tests, including disaster recovery and network stress tests

5.

Gather vendor monitoring dashboards (ensures clarity, visibility, and automated alerts in one place)

6.

Ensure a repeatable pipeline for evidence collection so that you have auto-generation of test execution logs and incident response metrics

The focus:

  • Your ability to restore a critical service within the defined recovery objective

  • Whether you can create a complete incident report (at least, placeholders) within regulatory timelines

  • Whether you record vendor SLAs. How exactly, if yes

What you get as a result:

  • Control test reports

  • Drill results with remediation tracking

  • Vendor monitoring logs

  • Structured evidence repository

Documentation reflects tested reality and ensures proper audit.

Phase 4: Continuous testing and robust evidence (ongoingly)

Objective: Maintain everything you managed to achieve by this phase, review all of this constantly, and fine-tune if needed.

Typical Key Performance/Risk Indicators:

  • Backup restore success rate

  • RTO/RPO achievement rate

  • Incident mean time to detect (MTTD)

  • Incident mean time to resolve (MTTR)

  • Control pass rate in periodic testing

  • Percentage of critical vendors reviewed annually

Without exaggeration, supervisors can request information at any time. So, once again — focus on tested controls, traceable evidence, accountable owners, and board-level visibility.

Request a QA strategy session

Common failure points that teams underestimate

Probably, the most dangerous trap is a misperception of regulatory audits as paperwork. Rest assured, this is a 100% failure under the new rules. The most harmful misstep is clear, what else?

Diagram of DORA compliance framework showing six stages: initial audit, governance, risk identification, protection and anticipation, detection, and reaction and recovery, with supporting activities such as risk management, cybersecurity services, third-party risk management, and incident response.
DORA compliance framework and implementation stages

1. Policies exist only nominally. They don’t have proof

Your team wrote an excellent ICT risk policy, approved it with the board…and failed DORA compliance. Why? Because auditors don’t want texts, they want control software testing and proper evidence.

Possible pitfalls:

  • You update risk registers only before audits

  • You document change approvals in tickets, but don’t sample them

  • You conduct access reviews informally and don’t retain records

How to remediate:

1.

Put the documented results into a quarterly control sampling

2.

Link every policy logical section to a control and every control to evidence

3.

Assign named control owners with a concrete review cadence

If you can’t show timestamped proof, supervisors would say you don’t have control.

2. You have backups, yet don’t test restores constantly

Many fintech teams have backups. Yet, one of the DORA requirements is predictable restoration performance. For this reason, you need a testing trace.

Possible pitfalls:

  • You defined RTO and RPO targets, but never measured them

  • Restore tests executed only in lower environments

  • No proof that dependencies (databases, authentication, integrations) restore consistently

How to remediate:

1.

Schedule restore tests for critical services at least annually

2.

Record start and end times to validate RTO

3.

Document data integrity checks post-restore

4.

Track failed restore attempts as risk items until resolved

3. Incident classification is unclear, so reporting clocks get missed

The DORA framework also defines specific, time-bound reporting under ESA technical standards.

Possible pitfalls:

  • The regulation has specific “major incident” criteria, so if you don’t map your severity matrices to them, it’s a “-1 point”

  • All classifications should also be documented

  • No dry runs to test regulatory reporting timelines

How to remediate:

1.

Map business impact indicators directly to RTS thresholds

2.

Pre-fill reporting templates for common incident types

3.

Run simulation exercises

4.

Assign the owners of the regulatory submission

4. Vendor contracts don’t match DORA expectations, and you don’t have exit tactics

One of the most difficult DORA compliance articles (28-44). Yet, if we break them down, they don’t seem frightening.

Possible pitfalls:

  • You didn’t include compliance requirements, periodical check-ins, and exit tactics in contracts with third-party vendors

  • Even if you got all exit ducks in a row, you might not tested feasibility

How to remediate:

1.

Find DORA minimum requirements and include them in your contracts with vendors

2.

You already have the gradation of your business operations/functions. Now, cross-reference vendors using this gradation

3.

Estimate transition timelines in case of exit, data portability, and dependencies

4.

It’s always a good idea to invite critical vendors to join your crisis training

Supervisors usually want active and evidence-based oversight, just a text/document won’t close their audit.

DORA compliance checklist

We tried to include all essential elements of the real audit into this checklist. So, basically, you can conditionally understand whether you’d complete a real one successfully. Yet, the devil is in the details, so if you are not sure about any listed evidence for any item, prioritize the deep-dive into it.

DORA compliance checklist listing key requirements including ICT risk management framework, incident management, resilience testing, third-party risk management, information sharing, governance, documentation, and reporting timelines.
DORA compliance checklist for financial organizations

We grouped all items by pillars for your convenience.

Pillar 1. ICT risk management

  • Product Owner, CTO, and you identify and approve critical functions

  • Tech leader completes the asset inventory and assigns ownership

  • Tech + Law teams map the ICT risk register to services

  • You document change management controls

  • Teams conduct access reviews and retain the records

  • Dev and QA teams create a stored backup and restore test results

  • Teams introduce board-level ICT risk reporting

Evidence as proof:

  • Risk register export

  • Asset inventory snapshot

  • Restore test report (with timestamps!)

  • Access review log

  • Change approval samples

Pillar 2. ICT-related incident management and reporting

  • Incident classification matrix (aligned to RTS)

  • Reporting plan with timelines

  • Regulatory submission roles

  • You have an active post-incident review/AAR process

Evidence as proof:

  • Incident log with classification rationale

  • Submitted report copies

  • Internal escalation timeline records

  • Root cause analysis documents

Pillar 3. Digital operational resilience testing

  • Embed security and regression tests into daily builds

  • Schedule mandatory disaster recovery drills

  • Evidence as proof: Automated test execution logs, defect remediation tracking boards, penetration testing reports

Pillar 4. Third-party ICT risk management

  • Audit all vendor contracts for mandatory access and testing rights

  • Validate the technical feasibility of vendor exit strategies

  • Ensure DORA-aligned contract clauses in place

  • Ongoing monitor performance

Evidence as proof:

  • Vendor risk register

  • Contract clause mapping

  • SLA reports

  • Incident notification samples

  • Exit feasibility documentation

Pillar 5. Information sharing

  • Ensure you have secure channels for threat intelligence

  • Evidence as proof: Threat feed ingestion logs; internal security alerts triggered by external data

How DeviQA helps with DORA readiness

Policy and informational work with the team is just one of a complex system that helps ensure DORA compliance. A successful audit requires validation across engineering, testing, operations, and, in some sense, your operational/business maturity.

The DeviQA team treats DORA readiness as a delivery program, and our primary focus is the technical execution:

  • DORA gap assessment: Across all five pillars and overall maturity. We review your QA, security, and disaster recovery processes and develop 3 to 6 months remediation strategy.

  • Operational resilience testing: Two parts: (1) implementing measures so that your systems can survive and recover from real-world failures; and (2) properly documenting the results.

  • Security: End-to-end coordination of penetration testing, vulnerability scanning, IAM, authentication testing, and API security validation. Fine-tuning continuous regression testing.

  • Transiting to ongoing support: Your software and vendors will change. We cover that with regular resilience and recovery testing, DR and failover drills, and ongoing support during regulatory reviews.

A tested control environment, documented validation results, and a robust evidence set create confidence and strong theses your team can use during the audit.

In a strategic consultation, we can assess your current potential, giving your maturity, readiness, and specific processes.

Book a strategic QA consultation

Dmitry Reznik

About the author

Dmitry Reznik

Chief Technology Officer

Dmitry Reznik is the Chief Technology Officer and co-founder at DeviQA, bringing deep technical expertise across software architecture, implementation, and long-term system operation.