Explore critical topics shaping today’s SAP landscape—from digital transformation and cloud migration to cybersecurity and business intelligence. Each topic is curated to provide in-depth insights, best practices, and the latest trends that help SAP professionals lead with confidence.
Discover how SAP strategies and implementations vary across global markets. Our regional content brings localized insights, regulations, and case studies to help you navigate the unique demands of your geography.
Get industry-specific insights into how SAP is transforming sectors like manufacturing, retail, energy, and healthcare. From supply chain optimization to real-time analytics, discover what’s working in your vertical.
Dive into the most talked-about themes shaping the SAP ecosystem right now. From cross-industry innovations to region-spanning initiatives, explore curated collections that spotlight what’s trending and driving transformation across the SAP community.
Explore critical topics shaping today’s SAP landscape—from digital transformation and cloud migration to cybersecurity and business intelligence. Each topic is curated to provide in-depth insights, best practices, and the latest trends that help SAP professionals lead with confidence.
Discover how SAP strategies and implementations vary across global markets. Our regional content brings localized insights, regulations, and case studies to help you navigate the unique demands of your geography.
Get industry-specific insights into how SAP is transforming sectors like manufacturing, retail, energy, and healthcare. From supply chain optimization to real-time analytics, discover what’s working in your vertical.
Dive into the most talked-about themes shaping the SAP ecosystem right now. From cross-industry innovations to region-spanning initiatives, explore curated collections that spotlight what’s trending and driving transformation across the SAP community.
Data virtualization is evolving into test data automation, which encompasses the entire lifecycle of data management, from discovery to governance, making it essential for modern engineering teams focused on CI/CD and fast release cycles.
The shift from data virtualization to test data automation matters because it aligns with the need for real-time access to compliant, privacy-ensured, and appropriately shaped test data, impacting software development teams that require efficient, reliable data for testing purposes.
Adopting test data automation enhances developer experience by providing self-service access to data, crucial for rapid delivery pipelines, thus transforming how teams approach data management and testing in today's fast-paced tech environment.
Despite data virtualisation remaining a valuable technique in enterprise architecture, it no longer fully encompasses the comprehensive outcomes required by modern engineering teams, which necessitate a broader approach through test data automation that includes discovery, transformation, delivery, and governance.
The phrase data virtualisation still matters. It describes a real architectural capability and many vendors continue to use it. But for engineering teams building in CI/CD, ephemeral environments, and fast release cycles, the term no longer describes the whole job. It captures one technique. It does not capture the complete outcome teams now need.125
A useful way to position it: data virtualisation is not wrong. It is narrower than the modern requirement.
1) What “data virtualisation” originally solved
In enterprise data architecture, Denodo defines data virtualization as a single data-access layer built from logical or virtual representations of physical sources, enabling real-time access across heterogeneous systems.1 In testing, that concept evolved into test data virtualization: lightweight, on-demand copies intended to avoid the time, cost, and storage overhead of repeatedly cloning full environments.
Broadcom’s earlier CA Virtual Test Data Manager materials make that framing explicit: they describe virtual copies of test data, on demand, in seconds, with a minimal footprint and self-service access for testers.2 Delphix still uses similar language today, describing how virtualization can produce ephemeral copies and virtual databases while enabling refresh and rollback workflows.56
Figure 1. The older model centered on creating virtual copies faster and more cheaply than full physical clones. That was valuable, but it did not by itself automate privacy, subsetting, synthesis, and CI/CD delivery end to end.
2) Why that framing is too narrow now
Modern engineering teams do not just need access to a copy of data. They need the right data, in the right shape, with the right privacy controls, delivered at the right moment in the delivery workflow.
That broader requirement shows up clearly in current product documentation. Redgate describes test data management as the process of providing DevOps teams with test data, typically including copying production data, anonymization or masking, and sometimes virtualization.3 Tonic Structural emphasizes sensitive data detection, transformation, safe dataset creation, and CI/CD integration so generated data is available on demand.4 Delphix argues that modern teams must move past legacy TDM practices and automate data delivery to CI/CD pipelines via unified APIs.6
The privacy dimension matters too. NIST’s guidance on de-identification notes that organizations may choose models such as publishing synthetic data, and explicitly warns that tools which merely mask personal information may not provide sufficient functionality for de-identification.7 In other words, modern test data workflows are increasingly about governed data transformation, not only cloning or virtual access.
Old question: How do we virtualise test data faster?
Better modern question: How do we automate compliant, production-like test data delivery across every environment and pipeline?
3) The modern model: test data automation
Test data automation is the more useful umbrella term because it describes the full lifecycle. It covers not just how data is accessed, but how it is discovered, transformed, generated, provisioned, refreshed, audited, and reused.
Figure 2. The modern requirement is lifecycle automation: discovery, subsetting, de-identification, synthesis, API delivery, environment reset, and governance.
The capability stack
A practical modern stack usually includes:
Discovery and classification to find sensitive fields and map relationships.
Subsetting to create smaller, representative datasets while preserving referential integrity.
Masking or de-identification to protect sensitive data in non-production use.
Synthetic generation when real production-like data is unavailable, insufficient, or too risky.
Provisioning APIs and self-service workflows so developers, testers, and pipelines can request data without waiting on tickets.
Refresh, rewind, share, and tear-down controls for fast iteration in isolated environments.
Governance so teams know what data was provisioned, how it was transformed, and where it was used.
4) Data virtualisation vs. test data automation
Dimension
Data virtualisation
Test data automation
Primary focus
Fast, lightweight access to virtualized or logical copies of data.
End-to-end automation of discovering, preparing, delivering, and governing test data.
Core mechanism
Logical access or virtual copies.
Subsetting, masking, synthesis, APIs, orchestration, and policy controls.
Privacy model
Often handled separately or added later.
Built into the workflow through classification, de-identification, and synthetic options.
Developer experience
Faster than full clones, but may still depend on requests or shared-state management.
Self-service and API-driven, designed for on-demand provisioning in delivery pipelines.
Environment model
Often centered on copies of a database or dataset.
Designed for ephemeral, parallel, and isolated test environments.
Best framing today
A useful technique.
A broader operating model.
5) The practical takeaway
If your strategy is still described mainly as data virtualisation, you may be describing only one layer of the value stack. The language modern teams respond to is broader because their requirement is broader:
safe by default
available on demand
small enough to move fast
realistic enough to test properly
easy to refresh and reset
designed to fit pipelines, not tickets
That is why test data automation is the better modern framing. It captures the outcome engineering teams actually care about: trusted test data delivered continuously, not just virtual copies created quickly.
Does this mean data virtualisation is obsolete?
No. It remains a real and valuable capability, and some vendors continue to use the term actively. The point is that virtualization is now part of a broader automation model, rather than the whole story.15
Why include synthetic data in the comparison?
Because modern privacy and usability requirements often go beyond simple masking. NIST explicitly includes synthetic data among the models agencies may choose for sharing and warns that masking alone may be insufficient for de-identification.7
What is the most defensible positioning statement?
Data virtualisation is one capability; test data automation is the modern operating model. That framing is strong without overclaiming that the old term has disappeared entirely.