Skip to main content

We'd prefer it if you saw us at our best.

Pega.com is not optimized for Internet Explorer. For the optimal experience, please use:

Close Deprecation Notice

Data virtualization

introduction to data virtualization

What is data virtualization?

Data virtualization is a modern data management approach that creates a unified data access layer, allowing users to view, access, and analyze data from multiple sources without needing to know the physical location or technical format of that data. Think of it as creating a virtual layer that sits between your applications and various data sources, providing a seamless way to work with data regardless of where it's stored.

Why is data virtualization important?

Data virtualization has become increasingly important in today's rapidly evolving technology landscape. It allows organizations to streamline data access and integration without the prolonged and often costly process of physical data consolidation.

Benefits of data virtualization

  • Simplified data access: Users can access data without managing and maintaining the underlying technical complexities of multiple data sources.
  • Real-time data integration: Data can be accessed in real time without the need to physically move or copy it between systems.
  • Increased agility: Organizations can quickly adapt to changing data requirements without rebuilding entire data pipelines.
  • Reduced IT dependency: Business users can access the data they need with less reliance on IT specialists.
  • Faster insights: By streamlining data access, organizations can discover insights and make decisions more quickly.
benefits of data virtualization

How does data virtualization work?

Data virtualization simplifies data access by creating a layer between data sources and applications. It uses metadata to map operations and process queries efficiently. When a query is made, the system finds the right data sources, runs optimized sub-queries in parallel, and combines the results in real time without storing copies. A caching system helps speed up frequently used data.

Data virtualization is just one aspect of an effective legacy transformation program. As an example, as a part of Pega Blueprint™, Pega Live Data enables teams to abstract and integrate data from multiple sources, including legacy systems, without requiring direct modifications or complex migrations.

how data virtualization works

Go from idea to app in a flash with the power of Pega Blueprint™

Learn more

Key components of data virtualization

Data sources layer

The foundational level where the system connects to and interacts with the various underlying data repositories

Abstraction layer

Allows data from many differing sources to be viewed as if it was from one single source

Virtualization engine

Responsible for orchestrating the access, integration, and delivery of virtualized data

Data integration services

Provides a unified view of data without physically moving it

Consumption layer

Serves as the point of access for users and applications to interact with the virtualized data

Security & governance framework

Ensures that the virtualized data is accessed and used responsibly and securely

Use cases for data virtualization

Supply chain management

Standardizing data from diverse sources for seamless collaboration between suppliers and manufacturers

Application development

Allowing developers to access data from multiple sources through a single interface, as Pega does with Data Pages

Virtual data marts

Creating unified views of data from various sources for specific departments or functions

Getting started with data virtualization

Step 1

Conduct a thorough assessment of your current data infrastructure to identify the key business use cases that would benefit from improved data access and integration.

Step 2

Once the foundational analysis is complete, organizations should prioritize establishing a cross-functional team that will facilitate the effective design and implementation of the data virtualization layer.

Step 3

Develop a detailed roadmap, highlighting important milestones and metrics for tracking progress and measuring success.

Step 4

Continuous evaluation and adaptation are essential, as data virtualization is an iterative process that evolves alongside the organization's data strategy and technological advancements.

Data virtualization enables applications to operate independently of underlying data sources, boosting agility and efficiency. For faster optimization, especially in the context of a larger legacy transformation effort, an AI workflow builder like Pega Blueprint can help organizations visualize and quickly integrate new data workflows, without having to re-architect disparate systems.

Frequently asked questions on data virtualization

Data virtualization handles diverse data sources and formats by using connectors and adapters to access them, managing metadata to understand their structure, and creating an abstraction layer that presents a unified, logical view.

Data virtualization can be deployed:

  • On-premises
  • In the cloud
  • As a hybrid solution
  • Using containerization

Data virtualization can enhance performance by reducing data movement by querying at the source, optimizing queries to limit data retrieval, and enabling real-time access for time-sensitive applications. It also speeds up report development by reducing the need for extensive ETL (extract, transform, load) processes.

Data virtualization allows access to legacy data without migration – so businesses can evaluate and retire outdated systems while preserving valuable information. This strategy speeds up modernization, reduces technical debt, and supports innovation without disrupting business operations.

Pega's Live Data virtualization layer acts as an abstraction layer, meaning it hides the complexities of underlying data sources from the applications that use them. This is especially important when dealing with diverse and often outdated legacy systems. It also allows Pega Blueprint to create a logical data model that rationalizes integrations between workflows and back-end systems.

Ready to learn more?

Legacy modernization slide show

Tech knowledge

Learn about legacy modernization, the first step toward complete legacy transformation.
Workflow Optimization

Tech knowledge

Discover how workflow optimization can transform operations, reduce costs, and improve customer satisfaction.
Intelligent automation slideshow

Tech knowledge

Learn how intelligent automation helps enterprises build applications that connect people, processes, and technology across organizations.
Predictive analytics slideshow

Tech knowledge

Discover how predictive analytics predicts customer needs so businesses can create better customer engagement strategies.
AI application Development Carousel

Tech knowledge

Find out how AI gives enterprises a competitive edge by developing better apps at a faster rate.

Explore what’s possible with Pega

Try now
Share this page Share via X Share via LinkedIn Copying...