Blog

  • Backup Acceleration, Restore Reality, and Why Replication Matters for Business Continuity

    Backup Acceleration, Restore Reality, and Why Replication Matters for Business Continuity

    There’s a growing trend in the enterprise backup market where vendors are introducing “backup acceleration appliances.” When you strip away the marketing, most of these are essentially very fast NVMe landing zones that sit in front of slower backup storage. They act like a write cache for backups — ingest data quickly, then move it to slower storage later.

    These devices absolutely have a place, but it’s important to understand what problem they actually solve, and what they don’t.

    Backup Acceleration: Solving the Backup Window Problem

    In many environments, the biggest backup challenge is still the backup window. Large data volumes, high change rates, and slower deduplicated or object storage repositories mean backups can run into business hours or impact production systems.

    A fast NVMe landing zone solves this by allowing backup jobs to write at very high speed. The data is then de-staged to slower storage in the background after the backup job has completed. From the backup server’s perspective, the job finishes much sooner.

    This can shrink backup windows, reduce load on production systems, and prevent backups running into business hours. For backup ingestion performance, this is a very effective solution.

    The Important Reality: It Doesn’t Help Restore Speeds

    Where buyers need to be careful is assuming that faster backups automatically mean faster restores. In most architectures, once the backup data has been moved from the NVMe landing zone to the main backup repository, restores still have to come from that slower storage and often involve rehydrating compressed or deduplicated data.

    So while backup jobs finish faster, restore times are often unchanged.

    This highlights a critical point that is often misunderstood in backup design: backup performance and restore performance are completely different engineering problems. Backup acceleration appliances solve backup ingest bottlenecks, not recovery time objectives.

    Backup vs Business Continuity

    This leads to a bigger architectural issue: backups are often treated as the entire disaster recovery strategy. They are not. Backups are for retention, rollback, and protection. They are not designed for rapid recovery of entire platforms.

    If the business requirement is fast recovery and low downtime, replication needs to be part of the solution. Replication keeps a near-live copy of systems in another location so recovery becomes a failover operation rather than a full restore.

    The simplest way to think about it is:

    Backup = Data Protection
    Replication = Business Continuity

    You still need backups for retention, compliance, and ransomware recovery, but backups alone are not a business continuity strategy.

    The Real Takeaway

    Backup acceleration appliances are useful for shrinking backup windows and solving ingest performance problems, but they will not significantly improve restore speeds. If the real requirement is faster recovery, the discussion should move beyond backup storage and toward replication and recovery architecture.

    Because ultimately, the business does not care how fast backups run — they care how fast systems come back.

  • When Your Backup Software Talks to the Internet: AI, Veeam Intelligence, and the New Privacy Reality

    When Your Backup Software Talks to the Internet: AI, Veeam Intelligence, and the New Privacy Reality

    A recent discussion on the Veeam R&D forums highlighted something that is going to become increasingly common in enterprise infrastructure: AI features that require your infrastructure data to be sent somewhere else for processing.

    The thread in question discussed Veeam Intelligence, the AI assistant integrated into Veeam Backup & Replication and Veeam ONE.
    Forum thread:
    https://forums.veeam.com/veeam-backup-replication-f2/veeam-intelligence-t102502.html

    While the discussion was technical and fairly calm, it touches on a much bigger issue: AI fundamentally changes the data-flow model of technical operations tools.


    What the Forum Discussion Revealed

    The key takeaway from the discussion was that when Veeam Intelligence is enabled, data can be sent to Veeam servers, depending on the mode used.

    From the discussion and documentation:

    • Basic Mode
      • Sends only the text of the user’s query to Veeam servers.
      • Uses documentation and knowledge base content to generate answers.
    • Advanced Mode
      • Uses environment data via product APIs to generate environment-specific answers.
      • This can include information such as:
        • Backup jobs
        • Protected VMs
        • Storage and applications
        • Alarms and monitoring data
        • Infrastructure information

    This behaviour is also documented in Veeam’s own documentation: the AI assistant can access infrastructure data via APIs to generate environment-aware responses.

    In other words, the AI assistant is not just answering documentation questions — it may be analyzing your backup environment by sending API output to a cloud AI service.

    That is a very different operational model from traditional backup software.


    The Important Line From the Forum

    One of the most telling responses in the thread was essentially:

    If in doubt: don’t use it.

    That line sums up the current state of AI features in infrastructure software surprisingly well.

    Not because vendors are doing anything malicious — but because the data flows are complex, evolving, and often not fully transparent to administrators.


    AI Changes the Trust Boundary

    Historically, backup software had a very clear trust model:

    • Backup server
    • Backup storage
    • Tape / offsite
    • Admin console
    • Everything inside your network

    AI assistants break that model.

    Now the workflow may look like this:

    1. Admin asks a question
    2. Software queries internal APIs
    3. API output is packaged
    4. Data is sent over HTTPS to vendor AI service
    5. AI processes it
    6. Response returned

    Even if data is anonymised or redacted, the backup environment metadata itself is extremely sensitive:

    • Server names
    • Job names
    • VM names
    • Network structure
    • Storage layout
    • Alarm data
    • Backup failures
    • Retention policies
    • Repository locations

    From a security perspective, this is effectively a blueprint of your disaster recovery strategy leaving your network.


    The Bigger Issue: AI in Technical Operations

    This is not about Veeam specifically.
    This is about AI in operational tooling.

    The same pattern is now appearing in:

    • Backup software
    • Monitoring platforms
    • SIEM tools
    • DevOps platforms
    • Cloud management tools
    • Documentation platforms
    • Ticketing systems
    • Code repositories
    • Network monitoring tools

    AI assistants work best when they have context, and context means data, and data means data leaving your environment.

    This creates several new challenges:

    1. Data Sovereignty

    Where is the AI processing happening?
    Which country?
    Which legal jurisdiction?

    2. Data Retention

    Is the data stored?
    For how long?
    Is it used to train models?

    3. Change Over Time

    One forum comment pointed out something very important:
    What is sent today may change in a future update.

    That is a huge operational governance problem.

    4. Invisible Data Flows

    Traditional tools:

    • You know when logs are exported
    • You know when backups are copied
    • You know when replication occurs

    AI tools:

    • Data flows when someone asks a question
    • Data flows when AI analyzes logs
    • Data flows when AI generates recommendations
    • Data flows in the background

    This is a completely new operational risk category.


    The New Question for Infrastructure Teams

    We used to ask:

    Is this tool secure?

    Now we have to ask:

    What data leaves our environment when this tool uses AI?

    That is a very different question.

    And most organizations do not yet have governance policies for this.


    Practical Recommendations for Organizations

    If you are running enterprise infrastructure and AI features are appearing in your tools, you probably need new policies:

    1. Treat AI assistants like external services.
    2. Review what data is transmitted.
    3. Check retention policies.
    4. Decide whether AI features are allowed in production environments.
    5. Consider enabling AI only in test environments.
    6. Monitor outbound connections from infrastructure servers.
    7. Include AI data flows in security reviews.
    8. Document AI features in risk registers.
    9. Inform compliance and legal teams.
    10. Assume AI features will expand over time.

    The Bigger Picture

    The Veeam forum discussion is interesting not because of Veeam specifically, but because it shows something important:

    AI is quietly changing how infrastructure software works.

    Not in the UI.
    Not in the features.
    But in where your operational data goes.

    For the last 30 years, infrastructure tools mostly ran inside your network.

    AI tools do not.

    And that is going to be one of the biggest operational and security shifts in enterprise IT over the next decade.


    Final Thought

    Backups are supposed to be the last line of defence — the one system that absolutely must be isolated, controlled, and secure.

    If AI assistants connected to backup infrastructure start exporting environment metadata to external services, even for legitimate reasons, then organizations need to think very carefully about where the new trust boundary actually is.

    Because with AI-enabled infrastructure tools, your backup environment may no longer be entirely inside your network — even if the servers are.

  • Why Ransomware Recovery Isn’t Just Disaster Recovery

    For years, organisations have invested in disaster recovery (DR) capabilities to protect against outages, hardware failures, and natural disasters. These scenarios are well understood: systems fail, recovery plans are executed, and services are restored as quickly as possible.

    But ransomware changes the game entirely.

    While ransomware incidents are often initially treated like a traditional DR event—systems unavailable, data inaccessible—the reality is far more complex. Recovery from ransomware is not just a technical exercise; it becomes a multi-dimensional crisis involving legal, regulatory, financial, and even criminal considerations.

    Let’s unpack how ransomware recovery fundamentally differs from routine disaster recovery, and why that distinction matters.

    Trust Is Broken—Not Just Systems

    In a typical DR scenario, the assumption is that your backups and infrastructure are trustworthy. A storage array fails, a data centre goes offline, or a configuration error corrupts production systems—but your backups remain clean and reliable.

    With ransomware, that assumption disappears.

    You must now answer difficult questions:

    • Are the backups compromised?
    • Has data been exfiltrated?
    • Is the attacker still present in the environment?

    This forces a shift from recovery to forensic validation. Systems cannot simply be restored—they must be verified as safe.

    The Recovery Timeline Is No Longer Linear

    Routine DR follows a relatively predictable flow:

    1. Detect failure
    2. Failover or restore
    3. Resume operations

    Ransomware recovery is far less linear:

    • Detection may occur long after initial compromise
    • Containment must happen before recovery
    • Forensics and investigation run in parallel with restoration
    • Legal and regulatory notifications may delay action

    In some cases, organisations deliberately delay recovery to preserve evidence or prevent reinfection.

    New Stakeholders Enter the Room

    Perhaps the most significant difference is the sudden expansion of stakeholders involved in the recovery process.

    Law Enforcement

    Ransomware is a criminal act. Engaging law enforcement introduces:

    • Evidence preservation requirements
    • Guidance on threat actor behaviour
    • Potential involvement in broader investigations

    Critically, this can go beyond advisory involvement. In some cases, law enforcement may:

    • Require systems to remain untouched for evidentiary purposes
    • Mandate the creation of forensically sound copies before remediation
    • In extreme scenarios, impound physical hardware as part of an investigation

    This can significantly delay recovery efforts and create tension between operational urgency and investigative integrity.

    Data Protection Regulators

    If personal data is involved, regulatory obligations (e.g. breach notification timelines) come into play:

    • Was data accessed or exfiltrated?
    • When did the breach occur?
    • How quickly must authorities be notified?

    These questions often must be answered before recovery is fully underway.

    Cyber Insurance Providers

    Cyber insurers are now deeply embedded in ransomware response:

    • They may mandate use of specific incident response firms
    • Approval may be required before engaging vendors or incurring costs
    • Decisions around ransom payment (if considered) often involve insurers

    In many cases, insurers will insist that their own appointed incident response teams take control of the situation. While these teams are experienced, their primary objective is often to minimise the insurer’s financial exposure.

    This can create misalignment with the business, whose priorities are more likely to include:

    • Minimising operational disruption
    • Protecting customer trust
    • Restoring services as quickly as safely possible

    These differing objectives can lead to friction around key decisions such as recovery approach, vendor selection, and timelines.

    Legal and Communications Teams

    Internal stakeholders also expand:

    • Legal teams advise on liability and disclosure
    • PR/communications teams manage customer and media messaging
    • Executive leadership becomes directly involved much earlier

    Recovery becomes a board-level issue, not just an IT operation.

    Clean Recovery Is More Important Than Fast Recovery

    In traditional DR, speed is king. Meeting RTOs and RPOs is the primary objective.

    In ransomware scenarios, clean recovery outweighs fast recovery.

    Restoring infected systems too quickly can:

    • Reintroduce malware
    • Trigger repeat encryption
    • Undermine forensic investigations

    As a result, recovery often involves:

    • Rebuilding systems from bare metal
    • Rotating credentials and secrets
    • Segmenting networks before bringing services online

    This can significantly extend recovery timelines compared to standard DR expectations.

    Backups Become Strategic Assets—Or Liabilities

    Backups are the cornerstone of DR. In ransomware, they are both:

    • The primary path to recovery
    • A potential attack vector

    Modern ransomware operators actively target backup systems:

    • Deleting or encrypting backup repositories
    • Compromising backup credentials
    • Exploiting immutability gaps

    This elevates the importance of:

    • Air-gapped or immutable backups
    • Credential isolation
    • Backup system monitoring and hardening

    In ransomware recovery, not all backups are equal—some may be unusable or unsafe.

    Decision-Making Becomes More Complex

    Routine DR decisions are largely operational:

    • Which site to fail over to
    • Which backup to restore from

    Ransomware introduces ethical, legal, and financial dilemmas:

    • Should a ransom be paid?
    • What are the regulatory implications of payment?
    • What is the reputational impact of disclosure?

    These decisions require input from executives, legal counsel, insurers, and sometimes government agencies—not just IT.

    Testing for Ransomware Is Not the Same as Testing DR

    Many organisations confidently state they “test DR regularly.” However, ransomware exposes a gap: DR tests rarely simulate adversarial conditions.

    Effective ransomware readiness requires:

    • Testing recovery from known-good points in time
    • Validating backup immutability
    • Practicing full environment rebuilds
    • Running cross-functional incident response exercises

    This is closer to crisis simulation than traditional DR testing.

    Final Thoughts

    Ransomware recovery is not an extension of disaster recovery—it is a fundamentally different discipline.

    The involvement of external stakeholders such as law enforcement, regulators, and insurers transforms what was once a technical recovery process into a coordinated organisational response. Success depends not just on infrastructure and backups, but on governance, communication, and decision-making under pressure.

    For technology leaders, the implication is clear:
    If your recovery strategy assumes ransomware behaves like any other outage, it’s time to rethink your approach.

    Because when ransomware strikes, technical recovery is only one part of the problem.

  • Script of the day #1

    We had a requirement to create a report out of Veeam to list all backup jobs, contents, schedules, retention and any other config tweaks. It’s nothing you can’t get out of Veeam One, but not everyone uses it. Here’s what I came up with. It’s designed for Veeam v12.3 running on a local Windows system. The next version will support remote systems, including those running the new Linux software appliance.

    It’s a PowerShell script that you can run with no arguements to get a listing of all active jobs, or use “-IncludeDisabledJobs” to show disabled jobs also. You can also choose where the HTML report goes to with an “-OutputPath <path>” arguement