Active metadata management is a set of capabilities that enables continuous access and processing of metadata that support ongoing analysis over a different spectrum of maturity, use cases and vendor solutions. Active metadata outputs range from design recommendations based upon execution results and reports of runtime steps through, and indicators of, business outcomes achieved. The resulting recommendations from those analytics are issued as design inputs to humans or system-level instructions that are expected to have a response.
Gartner defines augmented data quality (ADQ) solutions as a set of capabilities for enhanced data quality experience aimed at improving insight discovery, next-best-action suggestions and process automation by leveraging AI/machine learning (ML) features, graph analysis and metadata analytics. Each of these technologies can work independently, or cooperatively, to create network effects that can be used to increase automation and effectiveness across a broad range of data quality use cases. These purpose-built solutions include a range of functions such as profiling and monitoring; data transformation; rule discovery and creation; matching, linking and merging; active metadata support; data remediation and role-based usability. These packaged solutions help implement and support the practice of data quality assurance, mostly embedded as part of a broader data and analytics (D&A) strategy. Various existing and upcoming use cases include: 1. Analytics, artificial intelligence and machine learning development 2. Data engineering 3. D&A governance 4. Master data management 5. Operational/transactional data quality
The data integration tools market comprises stand-alone software products that allow organizations to combine data from multiple sources, including performing tasks related to data access, transformation, enrichment and delivery. Data integration tools enable use cases such as data engineering, operational data integration, delivering modern data architectures, and enabling less-technical data integration. Data integration tools are procured by data and analytics (D&A) leaders and their teams for use by data engineers or less-technical users, such as business analysts or data scientists. These products are consumed as SaaS or deployed on-premises, in public or private cloud, or in hybrid configurations.
Data virtualization technology is based on the execution of distributed data management processing, primarily for queries, against multiple heterogeneous data sources, and federation of query results into virtual views. This is followed by the consumption of these virtual views by applications, query/reporting tools, message-oriented middleware or other data management infrastructure components. Data virtualization can be used to create virtualized and integrated views of data in-memory, rather than executing data movement and physically storing integrated views in a target data structure. It provides a layer of abstraction above the physical implementation of data, to simplify querying logic.