Gartner defines data integration as the discipline comprising the architectural patterns, methodologies and tools that allow organizations to achieve consistent access and delivery of data across a wide spectrum of data sources and data types to meet the data consumption requirements of business applications and end users. Data integration tools enable organizations to access, integrate, transform, process and move data that spans various endpoints and across any infrastructure to support their data integration use cases. The market for data integration tools includes vendors that offer a stand-alone software product (or products) to enable the construction and implementation of data access and data delivery infrastructure for a variety of data integration use cases.
Data masking is based on the premise that sensitive data can be transformed into less sensitive but still useful data. This is necessary to satisfy application testing use cases that require representative and coherent data, as well as analytics that involve the use of aggregate data for scoring, model building and statistical reporting. The market for data protection, DM included, continues to evolve with technologies designed to redact, anonymize, pseudonymize, or in some way deidentify data in order to protect it against confidentiality or privacy risk.
Data virtualization technology is based on the execution of distributed data management processing, primarily for queries, against multiple heterogeneous data sources, and federation of query results into virtual views. This is followed by the consumption of these virtual views by applications, query/reporting tools, message-oriented middleware or other data management infrastructure components. Data virtualization can be used to create virtualized and integrated views of data in-memory, rather than executing data movement and physically storing integrated views in a target data structure. It provides a layer of abstraction above the physical implementation of data, to simplify querying logic.
Elevating Test Data Management for DevOps is the process of providing DevOps teams with test data to evaluate the performance, and functionality of applications. This process typically includes copying production data, anonymization or masking, and, sometimes, virtualization. In some cases, specialized techniques, such as synthetic data generation, are appropriate. With that, it applies data masking techniques to protect sensitive data, including PII, PHI, PCI, and other corporate confidential information, from fraud and unauthorized access while preserving contextual meaning.