<?xml version="1.0" encoding="UTF-8" standalone="yes"?>
<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/">
    <channel>
        <copyright>Copyright TechTarget - All rights reserved</copyright>
        <description></description>
        <docs>https://cyber.law.harvard.edu/rss/rss.html</docs>
        <generator>Techtarget Feed Generator</generator>
        <language>en</language>
        <lastBuildDate>Tue, 28 Apr 2026 03:49:37 GMT</lastBuildDate>
        <link>https://www.techtarget.com/searchdatamanagement</link>
        <managingEditor>editor@techtarget.com</managingEditor>
        <item>
            <body>&lt;p&gt;Without effective &lt;a href="https://www.techtarget.com/searchdatamanagement/definition/data-governance"&gt;data governance&lt;/a&gt;, growing volumes of data in IT systems are likely to become a disorganized morass, limiting their potential use. The risk of data misuse also increases due to lax controls.&lt;/p&gt; 
&lt;p&gt;Conversely, well-governed data is consistent and accessible across the enterprise, enabling better-informed business decisions and more accurate analytics and AI applications. Companies are also less likely to experience serious data breaches or &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Top-3-data-privacy-challenges-and-how-to-address-them"&gt;data privacy issues&lt;/a&gt;, reducing their exposure to regulatory compliance problems, legal liabilities and reputational damage.&lt;/p&gt; 
&lt;p&gt;As a result, developing a data governance strategy is a &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Data-governance-responsibilities-now-belong-in-the-C-suite"&gt;high-priority item on the C-suite agenda&lt;/a&gt; in well-run companies. The chief data officer and other data leaders play central roles in that process, typically managing it and working closely with their business counterparts to create and then implement the governance strategy.&lt;/p&gt; 
&lt;p&gt;Getting started with data governance is a big undertaking that commonly requires a substantial budget and significant resource commitments. It might be tempting to buy a strategy from a consultancy or a &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/15-top-data-governance-tools-to-know-about"&gt;data governance software vendor&lt;/a&gt; that promises a packaged set of policies and tools. But for optimal alignment with your organization's business operations and processes, it's best to develop a tailored one in-house, using these seven steps.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="1. Document existing data governance processes"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;1. Document existing data governance processes&lt;/h2&gt;
 &lt;p&gt;Your company likely already has some data governance processes that should be incorporated into a formal strategy or replaced with new ones. Various people manage and oversee corporate data -- database administrators, backup admins, data architects and data quality analysts, for example. Document this by creating a directory of data assets and a corresponding list of managers and staff who are responsible or accountable for data.&lt;/p&gt;
 &lt;p&gt;Don't be surprised if this exercise reveals some sobering, even shocking, oversights and gaps. The existing informal approach might reflect a messy reality, but getting a picture of current processes sets the stage for establishing a more strategic data governance program.&lt;/p&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="2. Secure executive sponsorship for the data governance program"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;2. Secure executive sponsorship for the data governance program&lt;/h2&gt;
 &lt;p&gt;Enlist senior business executives to sponsor, fund and promote the governance program. Their buy-in and top-down influence are critical because effective data governance requires participation and cooperation by departments and business units across the enterprise. But how can you win executive support for an initiative that might not show clear bottom-line benefits, at least not right away?&lt;/p&gt;
 &lt;p&gt;Invoking fear, uncertainty and doubt is the usual default method. Horror stories of inaccurate data leading to bad business decisions, or of fines for failing to comply with &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Data-governance-regulations-that-executives-should-know"&gt;data privacy and protection laws&lt;/a&gt;, might be enough to convince business leaders to back a governance initiative. By itself, however, this defensive approach isn't the optimal way to secure long-term data governance commitments.&lt;/p&gt;
 &lt;p&gt;Instead, combine it with a more forward-looking approach. Explain that data governance is largely informal now and that the company needs a framework with more defined processes. Emphasize that implementing one will not only help &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Proactive-practices-for-data-quality-improvement"&gt;improve data quality&lt;/a&gt; and meet regulatory requirements but also make the organization more functional and resilient.&lt;/p&gt;
 &lt;p&gt;Also, address upfront an issue that often causes resentment among business stakeholders and users: the perception that data governance stifles creative uses of data. Data governance policies don't restrict innovation. It's quite the opposite: By creating a more reliable data foundation, effective governance enables new ideas to flourish while reducing the risk of improper data use.&lt;/p&gt;
 &lt;figure class="main-article-image full-col" data-img-fullsize="https://www.techtarget.com/rms/onlineImages/data_management-need_to_govern_data-f.png"&gt;
  &lt;img data-src="https://www.techtarget.com/rms/onlineImages/data_management-need_to_govern_data-f_mobile.png" class="lazy" data-srcset="https://www.techtarget.com/rms/onlineImages/data_management-need_to_govern_data-f_mobile.png 960w,https://www.techtarget.com/rms/onlineImages/data_management-need_to_govern_data-f.png 1280w" alt="Visual that lists key reasons why organizations need a data governance program." height="283" width="560"&gt;
  &lt;figcaption&gt;
   &lt;i class="icon pictures" data-icon="z"&gt;&lt;/i&gt;These are key reasons why organizations need to develop and implement an effective data governance strategy.
  &lt;/figcaption&gt;
  &lt;div class="main-article-image-enlarge"&gt;
   &lt;i class="icon" data-icon="w"&gt;&lt;/i&gt;
  &lt;/div&gt;
 &lt;/figure&gt;
&lt;/section&gt;      
&lt;section class="section main-article-chapter" data-menu-title="3. Improve data literacy and skills across the organization"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;3. Improve data literacy and skills across the organization&lt;/h2&gt;
 &lt;p&gt;End users who understand the potential value of data and how to use it effectively are more likely to recognize the need to protect data assets and prevent misuse. To foster that understanding across the enterprise, develop &lt;a href="https://www.techtarget.com/searchbusinessanalytics/tip/Develop-a-data-literacy-program-to-fit-your-company-needs"&gt;training to improve data literacy and skills&lt;/a&gt; as part of the governance strategy.&lt;/p&gt;
 &lt;p&gt;Enhanced data literacy also helps raise up data governance efforts in another way. End users often create duplicate reports, dashboards, spreadsheets and even entire databases because they don't know how to find existing ones. A &lt;a href="https://www.techtarget.com/searchbusinessanalytics/feature/How-business-leaders-can-make-a-data-literate-culture-stick"&gt;data-literate culture&lt;/a&gt; is better equipped to discover and reuse such assets, increasing efficiency and consistency and reducing the risk of data errors. This, in turn, helps streamline governance tasks.&lt;/p&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="4. Create a virtual governance team at first, then formalize roles"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;4. Create a virtual governance team at first, then formalize roles&lt;/h2&gt;
 &lt;p&gt;It's too much to ask a company to reorganize upfront to improve data governance. Instead, start by constructing some organizational scaffolding around the existing ad hoc data structures. Identify the key roles currently involved in governance processes and create a virtual team to improve coordination and collaboration.&lt;/p&gt;
 &lt;p&gt;As governance becomes more formalized, new roles will emerge. A data governance manager or vice president commonly leads a team of data governance specialists who coordinate the program. Some business or IT workers will become data stewards, with direct responsibility for implementing governance policies in particular data sets. That can be a full-time or part-time role, depending on the organization's size and the complexity of its governance needs.&lt;/p&gt;
 &lt;p&gt;Establishing a data governance council or committee is also a must. It typically includes the following members:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Representatives from all departments and business units.&lt;/li&gt; 
  &lt;li&gt;IT, legal and compliance executives.&lt;/li&gt; 
  &lt;li&gt;Data stewards or others with data ownership responsibilities.&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;p&gt;The council sets data governance policies, creates common data standards, prioritizes governance projects and resolves data-related disputes, among other responsibilities. Having one ensures there's broad input on data governance controls and helps pave the way for enterprise-wide adoption of the governance strategy.&lt;/p&gt;
&lt;/section&gt;      
&lt;section class="section main-article-chapter" data-menu-title="5. Decide how to measure the governance program's effectiveness"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;5. Decide how to measure the governance program's effectiveness&lt;/h2&gt;
 &lt;p&gt;For a data governance program to gain and maintain support in an organization, it's crucial to measure its effectiveness -- and show how it benefits the company. But effective data governance might not tangibly affect business performance. It also isn't easy to calculate KPIs for reduced business risks, such as avoiding regulatory fines or reputational damage.&lt;/p&gt;
 &lt;p&gt;Instead, identify &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Data-governance-metrics-Data-quality-data-literacy-and-more"&gt;key data governance metrics to track&lt;/a&gt; and tie them to business benefits such as improved decision-making, optimized business processes and stronger privacy protections. For example, use metrics on data accuracy, completeness, consistency, timeliness and duplication to &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/6-dimensions-of-data-quality-boost-data-performance"&gt;monitor data quality levels and document improvements&lt;/a&gt; that make data more reliable for analytics and AI applications.&lt;/p&gt;
 &lt;p&gt;Metrics also help identify governance issues. Tracking how often users access data provides insight into whether it's being used effectively. Low usage might indicate a lack of awareness or accessibility. An increase in the overall number of analytics users is a marker of the governance program's success, but further training might be required if metrics also show new users are creating reports and dashboards that duplicate existing ones.&lt;/p&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="6. Prioritize data governance for new AI applications"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;6. Prioritize data governance for new AI applications&lt;/h2&gt;
 &lt;p&gt;Governing data for AI applications is now a key consideration for data leaders and governance teams. Data readiness is &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Experts-share-practices-to-overcome-AI-data-readiness"&gt;critical to successful deployments&lt;/a&gt; of machine learning, generative AI and agentic AI tools. Effective data governance ensures that AI models are built on a solid foundation of high-quality data and don't use it in ways that violate privacy and ethics policies.&lt;/p&gt;
 &lt;p&gt;For example, retrieval-augmented generation (RAG) frameworks pose specific data governance challenges in enterprise AI applications. RAG enables large language models (LLMs) to &lt;a href="https://www.techtarget.com/searchenterpriseai/opinion/How-RAG-unlocks-the-power-of-enterprise-data"&gt;directly draw from enterprise data&lt;/a&gt; -- it retrieves relevant documents or records from internal knowledge bases and uses them to generate responses to user queries.&lt;/p&gt;
 &lt;p&gt;From a governance perspective, successful RAG use requires not only accurate, up-to-date data but also &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Data-lineage-documentation-imperative-to-data-quality"&gt;data lineage documentation&lt;/a&gt; that traces an LLM's output back to the original data sources for explainability and performance auditing. Well-managed access control is also necessary. Define and enforce user permissions in the RAG framework to prevent end users from inadvertently seeing data they can't access in conventional analytics applications.&lt;/p&gt;
 &lt;p&gt;In addition to incorporating AI-related governance processes into your data governance strategy, &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Data-and-AI-governance-must-team-up-for-AI-to-succeed"&gt;align them with an AI governance program&lt;/a&gt; that monitors and controls AI deployments more broadly. Data governance and AI governance are separate functions, but they go hand in hand and should be tightly integrated.&lt;/p&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="7. Select technologies that fit the data governance strategy"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;7. Select technologies that fit the data governance strategy&lt;/h2&gt;
 &lt;p&gt;Various technologies can be used in data governance initiatives. Data governance software automates program management tasks, such as policy development, process documentation, data classification and workflow management. Data catalogs provide a &lt;a href="https://www.techtarget.com/searchdatamanagement/answer/What-steps-are-key-to-building-a-data-catalog"&gt;unified inventory of data assets&lt;/a&gt;, with built-in governance, data lineage and data curation features. Analytics catalogs help users find relevant dashboards, reports and data&amp;nbsp;visualizations and provide guidance on how to use them appropriately.&lt;/p&gt;
 &lt;p&gt;For data processing, newer data lakehouse architectures combine the raw data storage of a data lake with the structured, governed repository of a data warehouse. Collapsing the separation between those two platforms streamlines data management and governance work and provides a single system that supports BI, advanced analytics and AI applications.&lt;/p&gt;
 &lt;p&gt;But don't build a data governance strategy around specific technologies. Selecting tools that align with the strategy and support its goals will put the governance program on track to deliver the expected benefits, rather than running into a technology dead end that undermines the program. &amp;nbsp;&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Editor's note:&lt;/b&gt; &lt;i&gt;This article was updated in April 2026 for timeliness and to add new information.&lt;/i&gt;&lt;/p&gt;
 &lt;p&gt;&lt;em&gt;Donald Farmer is a data strategist with 30-plus years of experience, including as a product team leader at Microsoft and Qlik. He advises global clients on data, analytics, AI and innovation strategy, with expertise spanning from tech giants to startups.&lt;/em&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>A strong data governance strategy enables more effective data use and helps prevent financial, legal and reputational problems. Follow these steps to develop one.</description>
            <image>https://cdn.ttgtmedia.com/rms/onlineimages/legal_g1169668297.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/tip/6-key-steps-to-develop-a-data-governance-strategy</link>
            <pubDate>Fri, 24 Apr 2026 17:01:00 GMT</pubDate>
            <title>How to develop a data governance strategy: 7 key steps</title>
        </item>
        <item>
            <body>&lt;p&gt;Even with a governance program in place, organizations can still fall short.&lt;/p&gt; 
&lt;p&gt;Lack of data ownership is a widespread issue. Many organizations are finding that vast amounts of data lack an assigned owner. According to the &lt;a target="_blank" href="https://cpl.thalesgroup.com/about-us/newsroom/ai-the-new-insider-threat-facing-organizations" rel="noopener"&gt;2026 Data Threat Report&lt;/a&gt; from global technology company Thales, only 34% of organizations know where all their data is stored, and only 39% can fully classify their data.&lt;/p&gt; 
&lt;p&gt;Executives must identify unowned datasets within their organizations and assign ownership to ensure enterprise data is appropriately governed and secured. Not doing so puts the organization at financial and reputational risk.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="Common data ownership oversights"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Common data ownership oversights&lt;/h2&gt;
 &lt;p&gt;Quality data has become a critical asset for organizations, especially for &lt;a href="https://www.techtarget.com/searchenterpriseai/feature/9-data-quality-issues-that-can-sideline-AI-projects"&gt;automation, analytics and AI&lt;/a&gt;. Data owners, data stewards and chief data officers are responsible for ensuring that data meets established standards and &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Data-lineage-documentation-imperative-to-data-quality"&gt;has trackable lineage&lt;/a&gt;.&lt;/p&gt;
 &lt;p&gt;But there are still types of data that organizations fail to capture and govern effectively, including:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;&lt;b&gt;Unstructured content from communication channels.&lt;/b&gt; Transcripts from messaging and collaboration apps, meeting transcripts and recordings, and content from emails and social media exchanges all lack ownership but might contain critical business information.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Data generated by shadow IT systems.&lt;/b&gt; About 70% of CIOs believe that the business units in their organizations deploy unsanctioned tech, according to Flexera's &lt;a href="https://www.flexera.com/resources/reports/ITV-REPORT-IT-Priorities" target="_blank" rel="noopener"&gt;2026 IT Priorities Report&lt;/a&gt;. That creates shadow data -- that is, data generated and stored outside the purview of the organization's IT and security controls as well as its data governance program.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Data developed in sandboxes and for temporary projects.&lt;/b&gt; This spans from developmental databases and test environments to intermediate datasets to &lt;a href="https://www.techtarget.com/searchenterpriseai/tip/Explore-the-role-of-training-data-in-AI-and-machine-learning"&gt;training data for AI and machine learning&lt;/a&gt; models.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Siloed data.&lt;/b&gt; Many data types fall into this category, including data stored in individual desktop folders, temporary reports, disconnected spreadsheets and data in legacy systems. Some also put vendor and third-party data in this category. Dark data also falls into this category, even though the business no longer uses it.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Operational and machine data.&lt;/b&gt; Various data types belong here, ranging from systems logs, IoT sensor data and API payloads to metadata. These datasets often lack a single owner and are overlooked in management and governance.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="Unowned data governance challenges"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Unowned data governance challenges&lt;/h2&gt;
 &lt;p&gt;Establishing governance and ownership over unowned data is challenging. Experts noted that these data types are hard to classify, easy to overlook and span multiple business functions. Consequently, no single business leader claims or is assigned ownership. Or it might be that data ownership lies with many or all leaders, which lacks accountability.&lt;/p&gt;
 &lt;p&gt;Regardless, unowned data puts the organization at risk of costly data breaches, &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Data-governance-regulations-that-executives-should-know"&gt;regulatory violations&lt;/a&gt; and inaccurate outputs from analytics, automation and AI initiatives due to poor data quality.&lt;/p&gt;
 &lt;p&gt;With such consequences in mind, many executives seek to improve their organization's overall data management and governance. According to Workvia's &lt;a target="_blank" href="https://www.workiva.com/sites/workiva/files/pdfs/workiva-2026-exec-benchmark-survey-en.pdf" rel="noopener"&gt;2026 Executive Benchmark Survey&lt;/a&gt;, business leaders ranked "strengthening data governance" as the second-highest priority for digital transformation projects, behind automating data collection and validation.&lt;/p&gt;
 &lt;p&gt;Similarly, Informatica's &lt;a href="https://www.informatica.com/about-us/news/news-releases/2026/01/20260127-new-global-cdo-report-reveals-data-governance-and-ai-literacy-as-key-accelerators-in-ai-adoption.html" target="_blank" rel="noopener"&gt;CDO Insights 2026 Report&lt;/a&gt; found that 86% of 600 surveyed data leaders planned to increase data management investments -- and 41% seek to boost data and AI governance. This is in response to concerns about poor data quality affecting business objectives.&lt;/p&gt;
 &lt;p&gt;Ultimately, data management and governance won't succeed if they can't be accounted for. Assigning ownership to unowned or over-shared data will help organizations strengthen data governance and boost data quality -- and their own trustworthiness.&lt;/p&gt;
 &lt;p&gt;&lt;em&gt;Mary K. Pratt is an award-winning freelance journalist with a focus on covering enterprise IT and cybersecurity management.&lt;/em&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>Organizations can't claim to have good data governance when they still have unowned data. Assigning ownership to siloed and dark data is critical to enterprise success.</description>
            <image>https://cdn.ttgtmedia.com/rms/onlineimages/folder-files07.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/tip/The-data-ownership-blind-spots-putting-organizations-at-risk</link>
            <pubDate>Fri, 24 Apr 2026 14:20:00 GMT</pubDate>
            <title>The data ownership blind spots putting organizations at risk</title>
        </item>
        <item>
            <body>&lt;p&gt;Data is one of an organization's most valuable assets. But without a comprehensive data strategy as a foundation, it often becomes fragmented, inconsistent and difficult to access or trust for business decision-making.&lt;/p&gt; 
&lt;p&gt;An effective enterprise data strategy establishes a structured approach to managing, governing and using data in alignment with business objectives. That enables companies to &lt;a href="https://www.techtarget.com/searchdatamanagement/opinion/Turning-data-into-a-strategic-advantage"&gt;unlock greater value from their data assets&lt;/a&gt; through improved decision-making, optimized business processes and increased operational efficiency. It also helps them boost innovation and gain a sustainable competitive advantage over less data-driven rivals.&lt;/p&gt; 
&lt;p&gt;The data strategy shouldn't focus on implementing new technologies. That comes later, driven by the strategy. Instead, it should set the direction for data management processes, address common data-related challenges and build the capabilities needed to &lt;a href="https://www.techtarget.com/searchdatamanagement/opinion/Data-management-and-governance-key-to-successful-AI-use"&gt;support planned data use&lt;/a&gt; across the enterprise.&lt;/p&gt; 
&lt;p&gt;Follow these 12 steps to develop a data strategy that accomplishes those things and positions your organization to realize long-term business benefits.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="1. Define clear business objectives for data initiatives"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;1. Define clear business objectives for data initiatives&lt;/h2&gt;
 &lt;p&gt;A successful enterprise data strategy is grounded in close alignment between data initiatives and business goals. Data management and analytics efforts should directly support priorities such as enabling better-informed decision-making, enhancing customer experience, improving business operations, &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Modern-data-architectures-as-a-risk-management-strategy"&gt;reducing risks&lt;/a&gt; and fostering innovation.&lt;/p&gt;
 &lt;p&gt;To achieve this alignment, work closely with senior executives and business managers to identify critical objectives that depend on effective data use. Engaging with key stakeholders at the outset ensures the data strategy addresses real business needs and guides appropriate technology choices to help meet them. Data initiatives tied to measurable business outcomes are more likely to gain executive support and sustained investment in the resources required for long-term success.&lt;/p&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="2. Assess the existing data landscape in your organization"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;2. Assess the existing data landscape in your organization&lt;/h2&gt;
 &lt;p&gt;Next, get a complete understanding of the organization's current data environment. A comprehensive assessment documents existing technologies, capabilities, challenges and opportunities for improvement. The data management team should conduct it with clear visibility across data domains and business processes enterprise-wide.&lt;/p&gt;
 &lt;p&gt;As part of the assessment, review source systems, data platforms, integration processes, governance structures and analytics applications, as well as how data flows between IT systems in different departments or business units. This uncovers issues such as data silos, inconsistent data definitions, limited metadata visibility and restricted access to relevant data. Identifying these gaps enables data leaders to prioritize initiatives and create a realistic roadmap for implementing the data strategy.&lt;/p&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="3. Specify the desired state for data management and analytics"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;3. Specify the desired state for data management and analytics&lt;/h2&gt;
 &lt;p&gt;Once the current-state assessment is complete and the results have been evaluated, articulate what works well and where changes are needed in data management and analytics processes. Defining the desired state clarifies what the organization can achieve through those changes. This vision should be based on the previously identified business imperatives for each data domain and function.&lt;/p&gt;
 &lt;p&gt;As part of this step, set data quality expectations and outline plans to harmonize core data management processes, such as data integration, metadata management and master data management. Doing so ensures consistency across systems and reliable access to &lt;a href="https://www.techtarget.com/searchdatamanagement/opinion/Trusted-data-is-the-foundation-of-data-driven-decisions-GenAI"&gt;relevant and trustworthy data&lt;/a&gt;.&lt;/p&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="4. Identify and prioritize critical data domains"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;4. Identify and prioritize critical data domains&lt;/h2&gt;
 &lt;p&gt;The strategic value of data varies. While an enterprise data strategy by definition should address all data domains, focus initial implementation efforts on the domains and associated data sets that are most critical to business operations and decision-making.&lt;/p&gt;
 &lt;p&gt;Identifying and prioritizing the highest‑value data domains enables data leaders to direct resources to areas where data management and analytics improvements will have the greatest business impact. In a retailer, for example, improving customer data quality enables more accurate analytics for targeted marketing and better customer service. Focusing on high‑value areas also helps demonstrate the data strategy's value and build momentum toward a more &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Use-these-steps-to-successfully-build-your-data-culture"&gt;data‑centric culture&lt;/a&gt;.&lt;/p&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="5. Create an implementation roadmap"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;5. Create an implementation roadmap&lt;/h2&gt;
 &lt;p&gt;After defining what your organization aims to achieve with data to support business priorities and what's required to do so, create an implementation roadmap that details how it will get there. A well-designed roadmap sequences data initiatives over time in a way that's achievable and measurable.&lt;/p&gt;
 &lt;p&gt;That requires balancing ambition with realism to enable sustained, disciplined progress on the data strategy rather than a series of disconnected short-term projects -- or, worse, overpromising on planned deployments. The roadmap should also connect long-term goals, such as becoming more data-driven or AI-enabled, to concrete steps across data management and analytics processes.&lt;/p&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="6. Develop data principles and strategic guardrails"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;6. Develop data principles and strategic guardrails&lt;/h2&gt;
 &lt;p&gt;Incorporate data principles and strategic guardrails into the data strategy so they actively shape decisions on data management and use, rather than being abstract guidelines. Foundational principles -- such as treating data as an enterprise asset, ensuring it's accurate and accessible, and establishing a single source of truth through transparent data management practices -- should directly inform the data operating model and architecture. This drives data consistency, reuse and trust across the organization.&lt;/p&gt;
 &lt;p&gt;Strategic guardrails are operational constraints and requirements in areas such as privacy, security, &lt;a href="https://www.techtarget.com/searchbusinessanalytics/feature/Why-ethical-use-of-data-is-so-important-to-enterprises"&gt;ethical data use&lt;/a&gt;, data quality and data platform design. Embed them in the data strategy as part of data governance policies and the implementation roadmap. Aligning suitable guardrails with the execution of data initiatives provides clear direction on appropriate data use, reduces data-related risks and enables BI, data science and business teams to innovate confidently within well-defined boundaries.&lt;/p&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="7. Build a data governance framework and assign data ownership"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;7. Build a data governance framework and assign data ownership&lt;/h2&gt;
 &lt;p&gt;A strong data governance program is a &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/6-key-components-of-a-successful-data-strategy"&gt;critical component of a data strategy&lt;/a&gt;. Effective data governance ensures that data remains consistent and reliable and that it's managed and used properly. Without it, various problems can arise. For example, different departments might create conflicting data definitions or data quality might deteriorate, compromising business decisions due to incomplete or inaccurate information.&lt;/p&gt;
 &lt;p&gt;Include &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/5-benefits-of-building-a-strong-data-governance-strategy"&gt;implementing the data governance framework&lt;/a&gt; as a foundational item in the data strategy's roadmap. The strategy should also detail expectations for managing data throughout its lifecycle and the role of data governance in supporting business objectives. Additionally, work with business stakeholders to assign ownership of data assets to appropriate individuals or teams and task them with ensuring the data they oversee is managed and used in accordance with governance policies.&lt;/p&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="8. Design an enterprise data architecture"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;8. Design an enterprise data architecture&lt;/h2&gt;
 &lt;p&gt;A &lt;a href="https://www.techtarget.com/searchdatamanagement/definition/What-is-data-architecture-A-data-management-blueprint"&gt;data architecture&lt;/a&gt; provides the technical foundation for managing and delivering data. It defines and visualizes how data is processed, integrated, stored and accessed across systems. However, in many organizations, the existing data architecture has been developed over time, often in a piecemeal fashion without an enterprise-wide focus. As a result, redundancies and gaps in the architecture create challenges with data access and use.&lt;/p&gt;
 &lt;p&gt;To address these issues, design an enterprise data architecture as part of the data strategy. In addition to a high-level architectural blueprint, it should include artifacts such as data models, data flow diagrams and documents that map data use to business processes. A well-designed data architecture guides data management processes, helps teams identify data challenges and supports both operational reporting and advanced analytics.&lt;/p&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="9. Implement security, privacy and regulatory compliance controls"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;9. Implement security, privacy and regulatory compliance controls&lt;/h2&gt;
 &lt;p&gt;Protecting the ever-increasing volumes of data that organizations collect and use is critical to avoiding business problems. In addition to strategic guardrails that set high-level boundaries on data management and use, a data strategy must include specific controls to &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Top-3-data-privacy-challenges-and-how-to-address-them"&gt;mitigate data security and privacy risks&lt;/a&gt;. For example, ensure that only authorized users can access sensitive data and that potential security threats can be detected and addressed quickly through predefined incident response plans.&lt;/p&gt;
 &lt;p&gt;Regulatory compliance is also a broader issue now due to the &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Data-governance-regulations-that-executives-should-know"&gt;growing number of data protection laws&lt;/a&gt; that require responsible management of personal information and transparency about how data is used. Integrate compliance mechanisms into the data strategy to help reduce legal risks and maintain trust with customers and business partners.&lt;/p&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="10. Enable data accessibility and increased data literacy"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;10. Enable data accessibility and increased data literacy&lt;/h2&gt;
 &lt;p&gt;Making trusted data accessible to the people who need it is a core objective of an enterprise data strategy. Data access is no longer restricted to technical specialists. A modern data strategy supports controlled, governed access for business users and data analysts through user-friendly dashboards, self-service analytics tools and &lt;a href="https://www.techtarget.com/searchdatamanagement/answer/What-steps-are-key-to-building-a-data-catalog"&gt;centralized data catalogs&lt;/a&gt;.&lt;/p&gt;
 &lt;p&gt;However, data accessibility alone isn't enough. Increased data literacy is also required across the organization to maximize the business value derived from data assets. As part of the data strategy, &lt;a href="https://www.techtarget.com/searchbusinessanalytics/tip/Develop-a-data-literacy-program-to-fit-your-company-needs"&gt;develop a data literacy program&lt;/a&gt; that sets expectations for workers and includes training to help them become more data-literate.&lt;/p&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="11. Build in support for BI, advanced analytics and AI applications"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;11. Build in support for BI, advanced analytics and AI applications&lt;/h2&gt;
 &lt;p&gt;In the past, data strategies often focused primarily on delivering data for use in BI and reporting applications. But now they must also focus on the data needed for expanding deployments of advanced analytics and AI applications in companies.&lt;/p&gt;
 &lt;p&gt;Build support for techniques and tools such as predictive analytics, machine learning and both generative AI and agentic AI into the enterprise data strategy. Used effectively, they help organizations identify patterns in large data sets, forecast trends, explore data more efficiently and optimize or automate business processes. However, they &lt;a href="https://www.techtarget.com/searchdatamanagement/opinion/The-future-of-AI-depends-on-better-data-not-bigger-models"&gt;depend on the strong data foundation&lt;/a&gt; that a data strategy provides.&lt;/p&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="12. Define metrics to track and evolve the data strategy"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;12. Define metrics to track and evolve the data strategy&lt;/h2&gt;
 &lt;p&gt;A data strategy should evolve over time as business priorities, data sets, technologies and regulations change -- and as problems are identified. To guide this evolution, define &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Data-governance-metrics-Data-quality-data-literacy-and-more"&gt;KPIs and other metrics&lt;/a&gt; to track the effectiveness of data initiatives. Include ones on things such as data quality, governance activities and data availability, security and use.&lt;/p&gt;
 &lt;p&gt;Monitoring them enables data and business leaders to evaluate progress on initiatives and identify areas for improvement. Regular reviews and continuous refinement of the data strategy ensure that it remains aligned with the organization's needs and continues to deliver business value. Spell out the need for that upfront, when setting expectations for the strategy, so it isn't a surprise to anyone.&lt;/p&gt;
 &lt;p&gt;&lt;em&gt;Anne Marie Smith, Ph.D., is an information management professional and consultant with broad experience across industries. She has also designed and delivered numerous data management courses and educational programs.&lt;/em&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>Here are 12 to-do items for data leaders developing a data strategy to help their organization use data more effectively for analytics and business decision-making.</description>
            <image>https://cdn.ttgtmedia.com/visuals/search400/iseries_database_manage/search400_article_014.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/tip/Developing-an-enterprise-data-strategy-10-steps-to-take</link>
            <pubDate>Fri, 17 Apr 2026 15:50:00 GMT</pubDate>
            <title>How to develop an enterprise data strategy: 12 key steps</title>
        </item>
        <item>
            <body>&lt;p&gt;Every business today is a data business. From the corner store tracking stock levels to the multinational manufacturer predicting market trends and shipping costs worldwide, all businesses run on data.&lt;/p&gt; 
&lt;p&gt;Specifically, they run on many types of data. For example, businesses of all kinds have transaction, reference and customer relationship data. We might also have industry-specific and external data, as well as metadata describing their formats and uses. Often, we integrate all these data types to create specialized analytics data sets. A well-planned data strategy keeps this complex ecosystem in order, with a &lt;a href="https://searchdatamanagement.techtarget.com/tip/5-principles-of-a-well-designed-data-architecture"&gt;strong data architecture&lt;/a&gt; as its foundation.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="Why do you need a data strategy?"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Why do you need a data strategy?&lt;/h2&gt;
 &lt;p&gt;A data strategy defines long-term objectives for how an organization uses data, along with the policies and practices that support them. To be successful, a data strategy must cover all data use cases – not just technical processes for &lt;a href="https://searchdatamanagement.techtarget.com/definition/data-management"&gt;data management&lt;/a&gt; and analytics, but also the human element.&lt;/p&gt;
 &lt;p&gt;No modern business can leave the management, security and use of such an important corporate asset to individual data architects or developers. A comprehensive data strategy, with broad involvement and support, ensures data is managed well and used effectively.&lt;/p&gt;
 &lt;p&gt;Data priorities differ across organizations, shaped by management strategies and business goals, so there's no generic template to follow. But there are six critical components every data strategy must include.&lt;/p&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="1. Data"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;1. Data&lt;/h2&gt;
 &lt;p&gt;This is the most fundamental component, of course. But all the advice that follows will be of no help if your data isn't safely stored and secured, well-maintained and ready for use. The strategic value of your data must be built on a solid base of enterprise data management. That includes integrating and processing your data, &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/6-dimensions-of-data-quality-boost-data-performance"&gt;validating its quality&lt;/a&gt;, governing its use and auditing the processes that affect it.&lt;/p&gt;
 &lt;p&gt;Once these basics are in place, I always recommend an &lt;a href="https://searchdatamanagement.techtarget.com/answer/What-steps-are-key-to-building-a-data-catalog"&gt;enterprise data catalog&lt;/a&gt; as a critical component of a data strategy. You can't strategize around data if you don't know what data you have. Data catalog tools are particularly useful for making data available to business users by providing detailed, descriptive metadata. Sometimes IT managers want to map their systems -- to know what data they have and where it resides. The IT team can create its own simplified data catalog for such needs.&lt;/p&gt;
 &lt;p&gt;The key questions are always the same. What data do I have? Where is it? Who can use it?&lt;/p&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="2. Tools"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;2. Tools&lt;/h2&gt;
 &lt;p&gt;Data catalog tools are provisioned by IT and data management teams who know how to use the various features in &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/16-top-data-catalog-software-tools-to-consider-using"&gt;data catalog software&lt;/a&gt;, set them up and deploy them. We can make a useful distinction between tools provided by IT and tools adopted by end users. Both play an important role in a data strategy, complementing rather than contradicting each other.&lt;/p&gt;
 &lt;p&gt;Data management tools are almost always the domain of IT. There are some lightweight data quality and data integration tools designed for business users, but data management remains largely a behind-the-scenes function.&lt;/p&gt;
 &lt;p&gt;IT often also deploys the BI tools used to create data visualizations, dashboards and reports. But data and business analysts might have their own preferences and choose different tools. That can work well so long as we put controls in place to &lt;a href="https://searchbusinessanalytics.techtarget.com/feature/Data-governance-framework-key-to-analytics-success"&gt;govern data access and usage&lt;/a&gt;. Likewise, data scientists might feel most comfortable using tools they already mastered or that support certain analytics methodologies.&lt;/p&gt;
 &lt;p&gt;In the past, most IT teams tried to prevent the use of unsanctioned, non-standard tools. Now, just as we've adapted to bring-your-own-device, analytics specialists commonly bring their own favored applications. A good data strategy embraces that diversity but with sensible limits. In this case, we can ask another question: What tools are appropriate to use? Enabling a data analyst to use a &lt;a href="https://searchbusinessanalytics.techtarget.com/definition/self-service-business-intelligence-BI"&gt;self-service BI&lt;/a&gt; application to build some dashboards is reasonable; allowing someone to build their own data warehouse beyond their skills and authority is not.&lt;/p&gt;
 &lt;figure class="main-article-image full-col" data-img-fullsize="https://www.techtarget.com/rms/onlineImages/data_management-key_stages_data_strategy-f.png"&gt;
  &lt;img data-src="https://www.techtarget.com/rms/onlineImages/data_management-key_stages_data_strategy-f_mobile.png" class="lazy" data-srcset="https://www.techtarget.com/rms/onlineImages/data_management-key_stages_data_strategy-f_mobile.png 960w,https://www.techtarget.com/rms/onlineImages/data_management-key_stages_data_strategy-f.png 1280w" alt="Key stages of the data strategy development process" height="267" width="560"&gt;
  &lt;figcaption&gt;
   &lt;i class="icon pictures" data-icon="z"&gt;&lt;/i&gt;These are the four main phases of developing a data strategy, according to Donna Burbank of Global Data Strategy.
  &lt;/figcaption&gt;
  &lt;div class="main-article-image-enlarge"&gt;
   &lt;i class="icon" data-icon="w"&gt;&lt;/i&gt;
  &lt;/div&gt;
 &lt;/figure&gt;
&lt;/section&gt;      
&lt;section class="section main-article-chapter" data-menu-title="3. Analytics techniques"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;3. Analytics techniques&lt;/h2&gt;
 &lt;p&gt;Just as we use various analytics tools depending on our needs, we also employ a variety of analytics techniques. Data visualization is a common example. We might also find uses for predictive analytics, text analytics, sentiment analysis and cluster analysis, to name a few advanced analytics techniques. They can be powerful and useful, but also need careful oversight. Without it, we might run afoul of &lt;a href="https://searchdatamanagement.techtarget.com/definition/data-governance"&gt;data governance&lt;/a&gt; and privacy laws.&lt;/p&gt;
 &lt;p&gt;Predictive analytics, for example, might show business value in optimizing equipment maintenance cycles. That's an uncontroversial use. But predictive techniques could also be used to help automate hiring or manage marketing promotions. In those cases, employees and consumers might have concerns about the reliability, fairness or openness of the process.&lt;/p&gt;
 &lt;p&gt;A data strategy must recognize that governing only data and tools might not suffice. We need to understand -- and train people to understand -- that not all analytics techniques are neutral. Some use cases, especially those involving personally identifiable information, won't be justified by their business value alone.&lt;/p&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="4. Collaboration"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;4. Collaboration&lt;/h2&gt;
 &lt;p&gt;In modern businesses, data use is typically more collaborative than in the past. Increased data literacy and easier-to-use tools mean more people can participate in analytics, as well as technical fields like &lt;a href="https://searchbusinessanalytics.techtarget.com/definition/data-preparation"&gt;data preparation&lt;/a&gt; and data quality.&lt;/p&gt;
 &lt;p&gt;Even closely controlled processes, such as data governance and primary data definition development, can be crowdsourced. For example, doing so can ensure that product names, error codes and managed processes reflect reality on the shop floor in a manufacturing company. Collaboration on primary data can also avoid that most frustrating customer service response: "There's no code for that."&lt;/p&gt;
 &lt;p&gt;Collaborative tools are also being used more, including file sharing, enterprise chat, messaging and video conferencing. Human beings are compulsive collaborators. We constantly share, discuss and debate with others. If collaboration isn't planned for, it will happen anyway -- unplanned.&lt;/p&gt;
 &lt;p&gt;Consider the role of data and analytics in your organization's business decisions and identify processes that involve engagement within and beyond teams. Use that insight to support the ability to share and comment on dashboards, reports and data visualizations.&lt;/p&gt;
 &lt;p&gt;For example, some &lt;a href="https://searchbusinessanalytics.techtarget.com/feature/How-to-evaluate-and-select-the-right-BI-analytics-tool"&gt;BI and analytics tools&lt;/a&gt; enable multiple users to annotate visualizations. Increasingly, they also integrate with chat and messaging apps. Even simple file sharing can be effective, especially when supported by enterprise-class scalability and security features.&lt;/p&gt;
&lt;/section&gt;      
&lt;section class="section main-article-chapter" data-menu-title="5. Documentation and auditing"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;5. Documentation and auditing&lt;/h2&gt;
 &lt;p&gt;In describing these data strategy components, I've emphasized the need to balance IT control with end users' freedom to do self-service when appropriate.&lt;/p&gt;
 &lt;p&gt;To find this balance, our strategic goals must be well documented. Successful data strategies are built on the ability to answer four questions about any element of the plan and any resource -- data, tools, etc. -- it incorporates.&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;What is appropriate?&lt;/li&gt; 
  &lt;li&gt;What is approved?&lt;/li&gt; 
  &lt;li&gt;What is the purpose?&lt;/li&gt; 
  &lt;li&gt;What is the governance policy?&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;p&gt;With good documentation of both the data strategy and the underlying data architecture, we can answer these questions before any new project or initiative. We should also be able to look back at any project and answer them retrospectively. By doing so, we put ourselves in a good position to audit how the data strategy is working. It can also help us assess compliance with data governance policies and other &lt;a href="https://searchdatamanagement.techtarget.com/feature/Data-model-design-tips-to-help-standardize-business-data"&gt;internal data standards&lt;/a&gt;.&lt;/p&gt;
 &lt;div class="youtube-wrapper"&gt;
  &lt;iframe width="560" height="315" src="https://www.youtube.com/embed/BqdPuwvwPk4?rel=0" title="YouTube video player" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture" allowfullscreen&gt;&lt;/iframe&gt;
 &lt;/div&gt;
&lt;/section&gt;      
&lt;section class="section main-article-chapter" data-menu-title="6. People"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;6. People&lt;/h2&gt;
 &lt;p&gt;The two most important elements of your data strategy are the bookends of this list: data and people. Organizations increasingly look for &lt;a href="https://www.techtarget.com/searchbusinessanalytics/tip/Data-literacy-training-requires-a-dual-approach"&gt;data literacy and analytics skills&lt;/a&gt; in new business hires. Almost every business school now teaches basic data analytics.&lt;/p&gt;
 &lt;p&gt;Data scientists remain in high demand, though the role has evolved significantly in recent years. AI and machine learning have reshaped what organizations require. The priority now is professionals who can not only analyze data, but also build and govern the systems that act on it.&lt;/p&gt;
 &lt;p&gt;You should also think carefully about IT and data management in your staffing and hiring. With so much technology running in the cloud and systems more robust than ever, it's tempting to think IT merely has to keep the lights on. It's not true. High availability, disaster recovery, meeting service-level agreements, supporting new business requirements and regulatory demands all fall into IT's domain.&lt;/p&gt;
 &lt;p&gt;Data architects, data integration developers, data engineers, database administrators and other &lt;a href="https://searchdatamanagement.techtarget.com/feature/Data-management-roles-Data-architect-vs-data-engineer-others"&gt;data management professionals&lt;/a&gt; also play key roles in meeting business needs. An IT staff that is savvy about the business is a great strategic advantage. That caliber of IT staff needs recognition and leadership support as much as any other role.&lt;/p&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="How to implement an effective data strategy"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;How to implement an effective data strategy&lt;/h2&gt;
 &lt;p&gt;These six key components aren't a complete guide to &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Developing-an-enterprise-data-strategy-10-steps-to-take"&gt;developing a data strategy&lt;/a&gt;. You also must consider broader concerns, such as budgets, competition, innovation, marketing plans, staffing policies and legal frameworks.&lt;/p&gt;
 &lt;p&gt;But you can &lt;a href="https://theodi.org/article/data-strategy-how-an-ecosystem-approach-can-help-shape-your-vision/"&gt;apply this thinking broadly&lt;/a&gt;. For example, your staffing plan could include guidelines for making better use of data and analytics based on strategic priorities. Product innovation is increasingly driven by data on customer feedback, user behavior and market trends.&lt;/p&gt;
 &lt;p&gt;Implementing a data strategy requires understanding your entire organization's strategic goals. From there, break down the role of data, how it will be managed and used and apply it consistently across production, finance, marketing and HR. The result will be a data strategy that is workable and flexible for ever-changing business pressures and needs.&lt;/p&gt;
 &lt;p&gt;&lt;strong&gt;Editor's note:&lt;/strong&gt;&amp;nbsp;&lt;em&gt;This article was republished in April 2026 to improve the reader experience.&amp;nbsp;&lt;/em&gt;&lt;/p&gt;
 &lt;p&gt;&lt;em&gt;Donald Farmer is a data strategist with 30+ years of experience, including as a product team leader at Microsoft and Qlik. He advises global clients on data, analytics, AI and innovation strategy, with expertise spanning from tech giants to startups. He lives in an experimental woodland home near Seattle.&lt;/em&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>These six elements are essential parts of an enterprise data strategy that will help meet business needs for information when paired with a solid data architecture.</description>
            <image>https://cdn.ttgtmedia.com/rms/onlineimages/strategy_a200792738.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/tip/6-key-components-of-a-successful-data-strategy</link>
            <pubDate>Wed, 15 Apr 2026 09:00:00 GMT</pubDate>
            <title>6 key components of a successful data strategy</title>
        </item>
        <item>
            <body>&lt;p&gt;AI made semantic search mainstream. Now, enterprise reality is forcing a strategic refinement.&lt;/p&gt; 
&lt;p&gt;Vector search became a common requirement after the release of ChatGPT and the rise of generative AI chatbots, and it's now a standard feature across many database platforms. But increasingly, vector search is no longer the sole decision point for data leaders looking for effective search and retrieval frameworks to support AI applications. As usage expands into business‑critical workflows, some implementations can strain under relevance gaps and rising operational and governance overhead. What matters now is hybrid search, which combines semantic similarity with keyword precision, because enterprise queries often require both meaning and exact terms. That shift is pushing many organizations to update their search approaches to match real business use.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="Why early vector search deployments break down"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Why early vector search deployments break down&lt;/h2&gt;
 &lt;p&gt;The practical problem is that many first-generation deployments struggle &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Assemble-the-layers-of-big-data-stack-architecture"&gt;as AI initiatives expand and data volumes grow&lt;/a&gt;. A system might handle similarity search reasonably well but stumble on terms specific to the business, such as product names, acronyms, customer identifiers, error codes and policy language. As those misses add up, it's time to reassess the type of search and retrieval architecture needed to fully support a more mature environment.&lt;/p&gt;
 &lt;p&gt;Hybrid search, also referred to as hybrid retrieval, sits at the center of these discussions because it reflects how enterprise search for AI applications works. Some queries depend on exact matches, &lt;a href="https://www.techtarget.com/searchdatamanagement/opinion/Why-data-semantics-matters-for-context-aware-systems"&gt;others on semantic similarity&lt;/a&gt;, and many require both. Hybrid search runs full-text and vector queries in parallel and blends the results into a single ranked list.&lt;/p&gt;
 &lt;p&gt;For database buyers, it's clear that hybrid search is the baseline. Standalone vector database products still have their place, and many also now support full-text search. But many teams can &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/AI-data-governance-guidance-that-gets-you-to-the-finish-line"&gt;store and query vector embeddings&lt;/a&gt; in their current systems, including core database engines and managed services. More and more, platform differentiation comes down to relevance, including filters that narrow results to the right scope and reranking capabilities that push the best candidates to the top of search results.&lt;/p&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="Three ways to modernize a retrieval framework"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Three ways to modernize a retrieval framework&lt;/h2&gt;
 &lt;p&gt;When a framework change is needed to help the business gain the expected benefits from AI applications, organizations have three options.&lt;/p&gt;
 &lt;h3&gt;1. Extend the existing platform&lt;/h3&gt;
 &lt;p&gt;The first path is to extend an existing data platform. This is usually the right move when vector-based retrieval is primarily an upgrade to the infrastructure the organization already uses and trusts.&lt;/p&gt;
 &lt;p&gt;The goal is to keep retrieval within the existing data stack while improving search support for AI workloads. MongoDB Atlas, Databricks, Snowflake, Azure Cosmos DB and PostgreSQL with the pgvector extension fit this pattern because vector search is &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Evaluating-the-different-types-of-DBMS-products"&gt;integrated into the broader platform&lt;/a&gt;, rather than deployed as a separate system.&lt;/p&gt;
 &lt;p&gt;For buyers, this path tends to make the most sense when governance continuity, platform simplicity and reusing the skills of existing operational teams matter more than introducing another specialized layer.&lt;/p&gt;
 &lt;h3&gt;2. Upgrade the search layer&lt;/h3&gt;
 &lt;p&gt;The second path is to upgrade the search layer. If most complaints focus on the search experience, the decision is less about the database and more about adopting a search-first layer optimized for relevance at scale.&lt;/p&gt;
 &lt;p&gt;Search-first platforms are typically designed around the idea that retrieval quality comes from combining full-text search and &lt;a href="https://www.techtarget.com/searchenterpriseai/tip/Top-RAG-tools"&gt;vector-based similarity search&lt;/a&gt; with ranking and filtering across indexed content. Azure AI Search, Elasticsearch, OpenSearch, Apache Solr and Algolia belong in this broader category.&lt;/p&gt;
 &lt;p&gt;This path is best when the enterprise needs stronger discovery, ranking and search quality across data sets, documents, knowledge bases, websites and other types of content.&amp;nbsp;&lt;/p&gt;
 &lt;h3&gt;3. Replace the existing vector platform or add a dedicated one&lt;/h3&gt;
 &lt;p&gt;The third path is to replace or add a specialized vector platform. This should usually be the escalation path, not the default.&lt;/p&gt;
 &lt;p&gt;While specialized platforms such as Pinecone, Weaviate, Qdrant and&amp;nbsp;Milvus (including the Zilliz Cloud service) also offer hybrid search capabilities, they are centered on dedicated vector retrieval infrastructure. That can make sense when retrieval has become strategic enough to justify a separate platform, or when current data and search environments no longer fit the workload.&lt;/p&gt;
 &lt;p&gt;A few use cases help clarify the three options. An enterprise building an internal knowledge assistant might not need a new platform if it can extend its existing data stack and improve search quality there. A company with a large digital content estate and weak search relevance might get more value from modernizing the search layer than from reworking the entire data platform. And a business that uses retrieval as a service across multiple AI products might decide it needs a dedicated vector platform. &amp;nbsp;&lt;/p&gt;
&lt;/section&gt;              
&lt;section class="section main-article-chapter" data-menu-title="Why relevance and governance steer retrieval choices"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Why relevance and governance steer retrieval choices&lt;/h2&gt;
 &lt;p&gt;Search and retrieval for AI are no longer just features that organizations set and forget. It is a core capability, and buying decisions now extend beyond whether a platform supports vector search.&lt;/p&gt;
 &lt;p&gt;For many organizations, the primary requirement is relevance quality. Can the platform support intent‑driven search while still returning precise results for specific business terms? Hybrid search has become the baseline by combining semantic understanding and keyword matching in a single request.&lt;/p&gt;
 &lt;p&gt;The second buying criterion is governance fit. Search and retrieval more frequently touch regulated, sensitive and business‑critical data. Can the platform work within the organization's governance model, rather than forcing new controls or workarounds?&lt;/p&gt;
 &lt;p&gt;The governance requirement is &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Data-governance-for-AI-requires-a-cross-functional-approach"&gt;only getting sharper as AI expands&lt;/a&gt;. A March 2026 Omdia &lt;a target="_blank" href="https://research.esg-global.com/reportaction/515202164/Toc" rel="noopener"&gt;report&lt;/a&gt; said 47% of 400 technical and business stakeholders cited data privacy as their organization's top risk in a generative AI initiative, and 38% underestimated security and governance costs. (Omdia is a division of Informa TechTarget.) Gartner's November 2025 report on cloud database management systems complements Omdia's findings, noting that metadata is emerging as the connective tissue for AI and search workflows.&lt;/p&gt;
 &lt;p&gt;As platforms move toward data fabrics and self-governing systems, integrated metadata becomes central to governance, observability and operational control, requiring a search and retrieval platform that improves over time and holds up under evolving AI workloads.&lt;/p&gt;
 &lt;p&gt;&lt;i&gt;Tom Walat is an editor and reporter for TechTarget, where he covers data technologies.&lt;/i&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>As AI workloads mature, enterprises face multiple data platform choices to improve search and retrieval capabilities while meeting governance and operational demands.</description>
            <image>https://cdn.ttgtmedia.com/rms/onlineimages/code_g684641103.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/feature/Hybrid-search-demands-reshape-retrieval-frameworks-for-AI</link>
            <pubDate>Tue, 14 Apr 2026 18:04:00 GMT</pubDate>
            <title>Hybrid search demands reshape retrieval frameworks for AI</title>
        </item>
        <item>
            <body>&lt;div&gt; 
 &lt;div&gt;&lt;/div&gt; 
&lt;/div&gt; 
&lt;p&gt;As regulatory landscapes evolve in an increasingly data-driven world, organizations face increasing pressure to ensure compliance.&lt;/p&gt; 
&lt;p&gt;Data-specific requirements govern &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Big-data-collection-processes-challenges-and-best-practices"&gt;how organizations collect, store, process and share data&lt;/a&gt;. Achieving compliance is an essential, ongoing activity that leadership must guide. To do so, executives must understand the various global regulatory requirements and the implications and risks of non-compliance. Successful compliance shows that the enterprise and its personnel fully embrace data governance across all data-related activities.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="Important data governance regulations"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Important data governance regulations&lt;/h2&gt;
 &lt;p&gt;&lt;a href="https://www.techtarget.com/searchdatamanagement/definition/data-governance"&gt;Data governance&lt;/a&gt; is an umbrella term encompassing several important activities, including data lifecycle, stewardship, security, privacy, destruction, quality, retention, access, classification and management. Identifying the specific enterprise data governance requirements under each regulation is essential for leadership, especially as their business expands internationally.&lt;/p&gt;
 &lt;p&gt;The following are data governance regulations that affect governance planning, strategies and procedures. Most of these laws apply to any organization doing business in the country, regardless of origin. &amp;nbsp;&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;&lt;b&gt;EU GDPR – &lt;/b&gt;A pivotal piece of data protection legislation, &lt;a href="https://www.techtarget.com/whatis/definition/General-Data-Protection-Regulation-GDPR"&gt;GDPR&lt;/a&gt; protects EU residents' personal data. It specifies data management strategies that organizations must follow, including conducting a data protection impact assessment to identify and address any risks. Failure to comply may result in significant financial penalties -- up to €20 million ($23 million) or 4% of the firm's worldwide annual revenue.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;CCPA – &lt;/b&gt;Consumers have the right to know how organizations collect and process their data. &lt;a href="https://www.techtarget.com/searchcio/definition/California-Consumer-Privacy-Act-CCPA"&gt;CCPA&lt;/a&gt; ensures California residents have the right to delete or limit the personal information organizations collect, to opt out of the sale of their data and correct inaccurate information. Violators can range from $2,663 (unintentional) to $7,988 (intentional).&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;UK GDPR and Data Protection Act – &lt;/b&gt;Enacted in 2018, &lt;a href="https://www.techtarget.com/searchdatabackup/definition/Data-Protection-Act-2018-DPA-2018"&gt;this legislation&lt;/a&gt; transposes the GDPR into UK law.&lt;b&gt; &lt;/b&gt;It requires strong data security, collection and processing practices. Penalties can range from £8.7 million ($11.5 million) to £17.5 million ($23.1 million), or 2% to 4% of the company's worldwide annual revenue -- whichever is higher.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;HIPAA – &lt;/b&gt;&lt;a href="https://www.techtarget.com/searchhealthit/definition/HIPAA"&gt;HIPAA Security and Privacy Rules&lt;/a&gt; apply specifically to the US healthcare system and govern rules on data access, security, use, and protected health information disclosures. It requires risk assessments and employee training. Violations are either civil or criminal, and penalties vary based on severity. Unknowing civil offenders face fines as low as $100 per violation, while willful offenders face fines up to $50,000 per violation. Criminal incidents can result in a fine of up to $250,000 and 10 years in prison.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;EU Data Governance Act – &lt;/b&gt;Launched in 2023, this legislation requires secure data sharing across the EU. It advocates data altruism, which examines how data can be used in the public interest. The act doesn't specify a blanket fine but offers criteria for determining penalties.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Sarbanes-Oxley Act (SOX) – &lt;/b&gt;&lt;a href="https://www.techtarget.com/searchcio/definition/Sarbanes-Oxley-Act"&gt;SOX legislation&lt;/a&gt; addresses issues in&lt;b&gt; &lt;/b&gt;financial management and reporting as applicable to all publicly traded companies in the US. It has strict controls on the accuracy, integrity, validation and verification of financial data. It also &lt;a href="https://www.techtarget.com/searchcio/definition/What-is-SOX-compliance-A-complete-guide-and-checklist"&gt;mandates effectiveness assessments&lt;/a&gt; for internal controls and data governance practices. Violators face 10 to 20 years in prison and hefty fines.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;UK Network and Information Systems regulations – &lt;/b&gt;These regulations focus on cybersecurity and incident reporting for network and information services providers. Cybersecurity requirements include regular security assessments and continuous improvements. Penalties cost up to £17 million ($22.4 million).&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Gramm-Leach-Bliley Act (GLBA) – &lt;/b&gt;&lt;a href="https://www.techtarget.com/searchcio/definition/Gramm-Leach-Bliley-Act"&gt;This US legislation&lt;/a&gt; mandates financial organizations establish information disclosure policies, implement security programs and perform regular risk assessments. Noncompliance can result in a $100,000 fine per violation.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Personal Information Protection Law (PIPL) –&lt;/b&gt; China's data protection law is among the toughest globally, applying to all enterprises handling personal data within China's borders. It has strict consent and trans-border data flow requirements. Penalties for non-compliance include ¥50 million RBM ($7 million), 5% of annual revenue or shutting down enterprises.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Digital Personal Data Protection Act (DPDPA) –&lt;/b&gt; India's 2023 act requires data fiduciaries to provide customers notices of their rights and inform them of the type of data they're collecting and why, with specific restrictions on cross-border data flows. The &lt;a href="https://www.techtarget.com/searchdatabackup/definition/Digital-Personal-Data-Protection-Act-2023"&gt;DPDPA&lt;/a&gt; mandates consent for any processing, with additional requirements regarding children's data. Penalties include up to ₹250 crore ($26.9 million).&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Personal Data Protection Act –&lt;/b&gt; Developed in Singapore, thi&lt;i&gt;s &lt;/i&gt;legislation is widely recognized throughout the Asia-Pacific region. It is consent-driven, mandates breach alerts and has retention limitations. If a company exceeds S$10 million ($7.7 million) in annual turnover in Singapore, it faces financial penalties up to 10% of that annual turnover. Otherwise, fines cannot exceed S$1 million ($778,000).&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Personal Data Protection Law –&lt;/b&gt; The UAE law regulates personal data processing, requiring consent and security, as well as strict rules for trans-border data flows. It gives individuals the right to correct inaccuracies and stop processing upon request. Noncompliance results in fines up to AED 5 million ($1.36 million).&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Law 0908 on Personal Data Protection – &lt;/b&gt;Morocco's legislation is one of Africa's most comprehensive data protection statutes. It requires organizations to register with the national government. Penalties for noncompliance include fines up to MAD 600,000 ($64,343) and/or imprisonment from three months to four years.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="Non-compliance risks for executives"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Non-compliance risks for executives&lt;/h2&gt;
 &lt;p paraeid="{2c1fe5c8-d26a-421b-895d-711c4f19c464}{16}" paraid="2110754361"&gt;While data governance is very much a technology-centered activity, it is also an executive accountability issue. If &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/5-benefits-of-building-a-strong-data-governance-strategy"&gt;data governance initiatives&lt;/a&gt; result in regulatory violations, improper AI use or data-related incidents, the highest levels of enterprise leadership -- including the C-suite and the board -- are liable. Penalties include fines, litigation, reputational damage and competitive risks.&lt;/p&gt;
&lt;/section&gt;  
&lt;section class="section main-article-chapter" data-menu-title="Compliance resources"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;&lt;iframe title="Risks of non-compliance with data regulations" aria-label="Table" id="datawrapper-chart-wuECR" src="https://datawrapper.dwcdn.net/wuECR/1/" scrolling="no" frameborder="0" style="width: 0; min-width: 100% !important; border: none;" height="819" data-external="1"&gt;&lt;/iframe&gt;&lt;span xml:lang="EN-US" data-contrast="none"&gt;&lt;span data-ccp-parastyle="heading 2"&gt;&lt;/span&gt;&lt;/span&gt;&lt;em&gt;&lt;/em&gt;Compliance resources&lt;/h2&gt;
 &lt;p&gt;Many organizations have created &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/5-data-governance-framework-examples"&gt;data governance frameworks&lt;/a&gt; that help enterprises establish data governance capabilities, including the following:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;&lt;b&gt;Data Management Body of Knowledge (DAMA-DMBOK) – &lt;/b&gt;Considered the industry standard for data governance, DAMA-DMBOK addresses data quality, stewardship and metadata, among other issues.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Control Objectives for Information and Related Technologies (COBIT) – &lt;/b&gt;Developed by ISACA, &lt;a href="https://www.techtarget.com/searchsecurity/definition/COBIT"&gt;COBIT&lt;/a&gt; offers strong controls and audit guidelines that align IT governance with business risk management and strategy.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;NIST Cybersecurity &amp;amp; Privacy Frameworks – &lt;/b&gt;NIST has two data governance frameworks: the&lt;b&gt; &lt;/b&gt;&lt;a href="https://www.techtarget.com/searchsecurity/definition/NIST-Cybersecurity-Framework"&gt;Cybersecurity Framework&lt;/a&gt; for reducing cybersecurity risks&lt;b&gt; &lt;/b&gt;and&lt;b&gt; &lt;/b&gt;the Privacy Framework to identify and manage privacy risks.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;ISO/IEC 38500 – &lt;/b&gt;Most recently updated in 2024, this standard is a key international standard for IT governance. It addresses legal, regulatory and ethical data use and provides vocabulary for IT governance.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Data Management Capability Assessment Model (DCAM) – &lt;/b&gt;Developed by the EDM Council, this framework defines a maturity model addressing data governance, quality and architecture.&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;p&gt;A &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/15-top-data-governance-tools-to-know-about"&gt;variety of tools and resources&lt;/a&gt; can help demonstrate compliance, including master data management tools, data discovery and classification tools, data catalogs and IAM systems. Senior management support and budget funding are essential for establishing a mature data governance program.&lt;/p&gt;
 &lt;p&gt;Consider investing in AI tools, which can greatly improve performance, provide better data analytics, automate repetitive processes and identify potential compliance issues. Existing tools and resources might have upgraded versions with AI capabilities.&lt;/p&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="How to achieve compliance"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;How to achieve compliance&lt;/h2&gt;
 &lt;p&gt;The following are best practices for executives to achieve optimal data governance compliance outcomes.&lt;/p&gt;
 &lt;h3&gt;Be accountable for and own data governance&lt;/h3&gt;
 &lt;p&gt;Just as organizations should have data owners and stewards for different domains, they should also &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Data-governance-responsibilities-now-belong-in-the-C-suite"&gt;make an executive responsible for data governance&lt;/a&gt; and compliance activities. Responsibilities include defining and measuring KPIs, conducting periodic board-level governance briefings and establishing partnerships with other departments, such as legal, HR, risk management and operations.&lt;/p&gt;
 &lt;h3&gt;Ensure that data governance is risk-based&lt;/h3&gt;
 &lt;p&gt;Establish data governance as a primary risk area. Add governance to a corporate risk register and examine risk from financial and regulatory perspectives. Map governance controls to appropriate regulations and frameworks. Building scenarios to address specific risk events, such as trans-border data violations, will help if they ever occur.&lt;/p&gt;
 &lt;h3&gt;Require auditable evidence on compliance activities&lt;/h3&gt;
 &lt;p&gt;Demonstrating data governance compliance at any time is essential in case of unannounced audits. Evidence of compliance includes project reports, compliance testing results, access management issues, &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Evaluating-data-quality-requires-clear-and-measurable-KPIs"&gt;data quality measurements&lt;/a&gt; and retention/deletion rules. Schedule quarterly audits for relevant controls and create evidence trails for regulator inquiries.&lt;/p&gt;
 &lt;h3&gt;Optimize data quality at the C-Level&lt;/h3&gt;
 &lt;p&gt;Data quality and lineage must be primary goals. Establish strong controls addressing &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/How-data-lineage-became-a-boardroom-metric"&gt;data quality, lineage and accuracy&lt;/a&gt;. Enforce data quality standards, launch quality checks and link metrics to business requirements. Establish beginning-to-end data lineage and ensure access to it.&lt;/p&gt;
 &lt;h3&gt;Enforce data access controls&lt;/h3&gt;
 &lt;p&gt;Senior leaders must ensure data access controls are consistently monitored, enforced and applied. Implement least privilege, role-based access controls, multi-factor authentication, segregation of duties and uninterrupted monitoring. Provide support for potential audits.&lt;/p&gt;
 &lt;h3&gt;Culture of compliance&lt;/h3&gt;
 &lt;p&gt;This starts with the C-suite and board. Mandate training for all employees on data-related activities and endorse data literacy throughout the enterprise. Regularly reiterate the importance of data governance at major company meetings. Support whistleblowing of any violations and note governance issues in performance reviews.&lt;/p&gt;
 &lt;p&gt;Linking all governance activities into a cohesive process also helps with compliance.&lt;b&gt; &lt;/b&gt;Data silos can spell disaster. Greater information sharing, along with the integration of security and privacy capabilities across systems, helps avoid this.&lt;/p&gt;
 &lt;h3&gt;Acquire technology that facilitates the compliance process&lt;/h3&gt;
 &lt;p&gt;The right technology ensures that governance activities are scalable, adaptable and automated. Automate data governance activities by integrating &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/AI-data-governance-guidance-that-gets-you-to-the-finish-line"&gt;AI tools&lt;/a&gt; with risk, privacy and security systems. However, be sure to provide &lt;a href="https://www.techtarget.com/searchenterpriseai/definition/AI-governance"&gt;AI governance&lt;/a&gt; oversight. When used correctly, it can facilitate the following data processes:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Reduce the likelihood of human error.&lt;/li&gt; 
  &lt;li&gt;Improve performance.&lt;/li&gt; 
  &lt;li&gt;Automate repetitive tasks such as data collection and classification.&lt;/li&gt; 
  &lt;li&gt;Identify potential compliance issues.&lt;/li&gt; 
  &lt;li&gt;Deliver reports for auditors.&lt;/li&gt; 
  &lt;li&gt;Ensure cross-border flow adheres to regulations.&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;h3&gt;Ensure that knowledge of regulatory activities is current&lt;/h3&gt;
 &lt;p&gt;Adapting to &lt;a href="https://www.techtarget.com/searchsecurity/tip/State-of-data-privacy-laws"&gt;regulatory changes&lt;/a&gt; and maintaining compliance are essential for enterprises. Executives should consistently monitor the global regulatory landscape. Review and assess regulatory changes, keep policies current and train governance teams to do the same.&lt;/p&gt;
 &lt;p&gt;&lt;em&gt;Paul Kirvan, FBCI, CISA, is an independent consultant and technical writer with more than 35 years of experience in business continuity, disaster recovery, resilience, cybersecurity, GRC, telecom and technical writing.&lt;/em&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>Growing national and international regulatory compliance demands aim to protect consumer data. Organizations must adhere to regulations or face noncompliance risks.</description>
            <image>https://cdn.ttgtmedia.com/rms/onlineimages/legal_g1065824400.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/tip/Data-governance-regulations-that-executives-should-know</link>
            <pubDate>Wed, 08 Apr 2026 15:31:00 GMT</pubDate>
            <title>Data governance regulations that executives should know</title>
        </item>
        <item>
            <body>&lt;p&gt;AI and cloud analytics applications are exposing a critical security gap for enterprises. While data is typically secured at rest and in transit, it often remains unprotected when being processed -- the time it is most actively used.&lt;/p&gt; 
&lt;p&gt;This gap has pushed data-in-use protection higher on the agenda for data leaders. Within the broader landscape of privacy‑enhancing technologies, &lt;a href="https://www.techtarget.com/searchsecurity/tip/Confidential-computing-use-cases-that-secure-data-in-use"&gt;confidential computing has emerged&lt;/a&gt; as the primary way to address this processing‑stage risk. It uses hardware‑isolated trusted execution environments (TEEs) to keep data encrypted during computation, enabling teams to expand AI workloads without overhauling data pipelines or weakening security.&lt;/p&gt; 
&lt;p&gt;Adoption trends suggest confidential computing is moving from a specialized control to a baseline expectation for AI and cloud analytics deployments. In a 2024 report, for example, Grand View Research&amp;nbsp;projected the global market for confidential computing would grow from an estimated $5.46 billion in 2023 to $153.8 billion by 2030, reflecting its increased role as a foundational component of data security.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="How data-in-use protection fits into existing pipelines"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;How data-in-use protection fits into existing pipelines&lt;/h2&gt;
 &lt;p&gt;Standard security leaves data vulnerable in system memory and CPUs. Data-in-use protection addresses this exposure problem by keeping information encrypted while workloads execute.&lt;/p&gt;
 &lt;p&gt;At the hardware layer, a &lt;a href="https://www.techtarget.com/searchitoperations/definition/trusted-execution-environment-TEE"&gt;TEE&lt;/a&gt; is a secure area that runs code and processes data independently from the rest of the system. It isolates data and processing operations to prevent unauthorized access. Even cloud administrators, host OSes and hypervisors do not have access to the data in a TEE.&lt;/p&gt;
 &lt;p&gt;Because confidential computing operates at the infrastructure layer, AI training and analytics jobs can often run in a TEE with minimal architectural changes. TEEs also transparently encrypt processing for applications, minimizing operational disruption while extending protection throughout the compute stage.&lt;/p&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="Compliance pressure moves into the processing layer"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Compliance pressure moves into the processing layer&lt;/h2&gt;
 &lt;p&gt;Rapidly evolving regulations are reshaping where organizations invest to secure AI and analytics workloads. A 2025 Stanford report found that &lt;a href="https://www.techtarget.com/searchenterpriseai/feature/AI-regulation-What-businesses-need-to-know"&gt;AI-related regulations&lt;/a&gt; issued by U.S. federal agencies more than doubled from 25 in 2023 to 59 in 2024. Similarly, the number of AI-related laws passed at the state level increased from 49 to 131.&lt;/p&gt;
 &lt;p&gt;Gartner predicts that by 2029, confidential computing will be used to secure more than 75% of processing operations running in shared infrastructure, such as public cloud services.&lt;/p&gt;
 &lt;p&gt;As sensitive data moves into AI pipelines, the pressure to document security grows. Processing-stage exposure is difficult to control and even harder to record without hardware-based locks. Audit teams and data governance functions that once focused only on storage encryption now require attestation that processing workloads run in protected environments.&lt;br&gt;&lt;br&gt;Several regulatory frameworks now explicitly require data-in-use protection:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;&lt;b&gt;EU AI Act.&lt;/b&gt; This new regulation requires documented data governance controls, including evidence of protection during &lt;a href="https://www.techtarget.com/searchenterpriseai/tip/10-steps-to-achieve-AI-implementation-in-your-business"&gt;all AI lifecycle stages&lt;/a&gt;.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;GDPR.&lt;/b&gt; Enforcement of the EU's data privacy regulation is expanding to include data in use, not just in storage or transit.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;PCI DSS v4.0.1.&lt;/b&gt; Requirements prevent sensitive authentication data from persisting in memory, such as RAM or memory dumps.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Digital Operational Resilience Act.&lt;/b&gt; DORA mandates data-in-use protection for major EU financial institutions, including controls on data handling within cloud and third‑party processing environments.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;NIST Cybersecurity Framework 2.0.&lt;/b&gt; Commonly known as CSF 2.0, it &lt;a target="_blank" href="https://www.nist.gov/cyberframework" rel="noopener"&gt;includes&lt;/a&gt; data-in-use protection within zero-trust security designs.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="What data leaders gain from data-in-use protection"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;What data leaders gain from data-in-use protection&lt;/h2&gt;
 &lt;p&gt;For data leaders, the value of confidential computing aligns with governance, legal and audit functions.&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;&lt;b&gt;Secure access to sensitive data. &lt;/b&gt;Healthcare records, financial transaction data, personally identifiable information and other regulated data often aren't used in AI and analytics initiatives due to processing risks. Confidential computing enables access to sensitive data sets without violating governance and security rules.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Reduced legal exposure&lt;/b&gt;. Confidential computing provides verifiable proof that sensitive workloads are processed in hardware-isolated environments, which is especially valuable for documenting regulatory compliance in third-party clouds.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Increased audit efficiency.&lt;/b&gt; The records of secure processing automatically produced by TEEs also reduce manual auditing work and improve audit verification on the use of sensitive data.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="Confidential computing use cases"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Confidential computing use cases&lt;/h2&gt;
 &lt;p&gt;The clearest proof of concept for confidential computing comes from highly regulated industries, where organizations face strict requirements around data handling, auditability and cross boundary data sharing.&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;&lt;b&gt;Healthcare. &lt;/b&gt;Hospitals and clinical research networks use confidential computing to support federated AI model training across institutions, keeping patient data private in shared systems or central databases.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Financial services.&lt;/b&gt; Banks and insurers use TEEs for fraud detection and risk modeling to reduce exposure when processing sensitive transaction data regulated by banking privacy rules.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Public sector.&lt;/b&gt; Agencies and partner organizations apply confidential computing to joint analytics projects without sharing raw data across organizational boundaries.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Telecom and IoT.&lt;/b&gt; Providers can use confidential computing to analyze customer and device data closer to the edge while limiting exposure during processing.&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;p&gt;Across these industries, common use cases include secure AI training, multi‑party analytics, &lt;a href="https://www.techtarget.com/searchenterpriseai/tip/How-to-navigate-data-sovereignty-for-AI-compliance"&gt;data sovereignty controls&lt;/a&gt;, and cloud backup and recovery workflows where restore operations can expose sensitive data.&lt;/p&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="How to evaluate data‑in‑use protection options"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;How to evaluate data‑in‑use protection options&lt;/h2&gt;
 &lt;p&gt;Approaches to data‑in‑use protection vary across cloud providers and the broader vendor ecosystem that includes data platforms, security and key management tools, and systems integrators. Before committing to a platform, data leaders should focus on proof points, regulatory alignment and how well it integrates with existing controls.&lt;/p&gt;
 &lt;p&gt;&lt;iframe title="" aria-label="Table" id="datawrapper-chart-KuZgD" src="https://datawrapper.dwcdn.net/KuZgD/1/" scrolling="no" frameborder="0" style="width: 0; min-width: 100% !important; border: none;" height="721" data-external="1"&gt;&lt;/iframe&gt;&lt;/p&gt;
 &lt;p&gt; &lt;script type="text/javascript"&gt;window.addEventListener("message",function(a){if(void 0!==a.data["datawrapper-height"]){var e=document.querySelectorAll("iframe");for(var t in a.data["datawrapper-height"])for(var r,i=0;r=e[i];i++)if(r.contentWindow===a.source){var d=a.data["datawrapper-height"][t]+"px";r.style.height=d}}});&lt;/script&gt; &lt;/p&gt;
 &lt;p&gt;&lt;i&gt;Sean Michael Kerner is an IT consultant, technology enthusiast and tinkerer. He has pulled Token Ring, configured NetWare and been known to compile his own Linux kernel. He consults with industry and media organizations on technology issues.&lt;/i&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>As sensitive data moves into AI pipelines, organizations must evaluate how to protect it during processing and what safeguards IT platforms provide for data in use.</description>
            <image>https://cdn.ttgtmedia.com/rms/onlineimages/security_a385093447.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/tip/AI-analytics-push-data-in-use-protection-up-priority-list</link>
            <pubDate>Tue, 07 Apr 2026 15:34:00 GMT</pubDate>
            <title>AI, analytics push data-in-use protection up priority list</title>
        </item>
        <item>
            <body>&lt;p&gt;Without properly prepared data, analytics and AI applications are unlikely to deliver the desired business outcomes. But &lt;a href="https://www.techtarget.com/searchbusinessanalytics/definition/data-preparation"&gt;data preparation&lt;/a&gt; is an inherently complex process that poses various challenges for data management and analytics teams.&lt;/p&gt; 
&lt;p&gt;Preparing data for planned uses requires substantial amounts of time and resources. Indeed, it typically accounts for most of the work involved in developing analytics applications. Large amounts of data in diverse formats &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Big-data-collection-processes-challenges-and-best-practices"&gt;collected from numerous sources&lt;/a&gt; must be combined and consolidated. The raw data routinely contains errors, anomalies, inconsistencies and other &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Proactive-practices-for-data-quality-improvement"&gt;data quality issues&lt;/a&gt;. Data sets might not include all the information an application requires. Conversely, some data might not be relevant to it.&lt;/p&gt; 
&lt;p&gt;Data preparation tools -- available as separate products or built into BI and data science platforms -- enable data scientists, data engineers, business analysts and other end users to prepare data themselves. However, these tools don't eliminate the challenges of data preparation. Data leaders must ensure users are sufficiently trained on the data prep process, including common challenges.&lt;/p&gt; 
&lt;p&gt;Effective data preparation also requires a multipronged approach. To aid self-service users, data quality analysts &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Data-cleansing-best-practices"&gt;profile and cleanse data&lt;/a&gt; upfront. Data integration developers run initial data transformation jobs. BI teams further transform, enrich and curate data sets for planned applications. They, too, must be prepared for the challenges of preparing data.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="7 top data preparation challenges"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;7 top data preparation challenges&lt;/h2&gt;
 &lt;p&gt;Because of its complexity, data preparation can't be left to chance. The following are seven notable challenges that disrupt efforts to create clean, consistent and complete data sets, along with advice on how to overcome each one.&lt;/p&gt;
 &lt;h3&gt;1. Inadequate or erroneous data profiling&lt;/h3&gt;
 &lt;p&gt;Data profiling should prevent end users from belatedly discovering data issues when running analytics applications -- or, worse, from having the analytics results be affected by faulty data they aren't aware of. But it might not do so due to the following scenarios:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Data team members or business users preparing data for a new application assume it's valid because it's already used in reports and dashboards. As a result, they don't fully profile the data. However, the existing uses masked underlying problems in the data set.&lt;/li&gt; 
  &lt;li&gt;Someone only profiles a sample data set from a large volume of data because of the time it would take to profile the full one. But the sampling approach doesn't detect anomalies and other issues in the full data set.&lt;/li&gt; 
  &lt;li&gt;Similarly, custom-coded SQL queries or spreadsheet functions used to profile data aren't comprehensive enough to find all the problems in the data.&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;p&gt;&lt;b&gt;How to overcome this challenge&lt;br&gt;&lt;/b&gt;Solid data profiling must be the starting point of the data preparation process. Data preparation tools can help: They include comprehensive functionality for profiling data sets in both source systems and the data platforms that analytics and AI applications run on.&lt;/p&gt;
 &lt;h3&gt;2. Missing or incomplete data&lt;/h3&gt;
 &lt;p&gt;Missing values and incomplete entries are common data quality issues. Examples include:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Null or blank fields.&lt;/li&gt; 
  &lt;li&gt;Zeros that represent a missing value rather than the number 0.&lt;/li&gt; 
  &lt;li&gt;Other types of placeholder values.&lt;/li&gt; 
  &lt;li&gt;Partial transaction records with missing details.&lt;/li&gt; 
  &lt;li&gt;Incomplete demographic data on customers.&lt;/li&gt; 
  &lt;li&gt;An entire field or row that's missing from a data set.&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;p&gt;Missing or incomplete data can adversely affect business decisions driven by analytics applications and create &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Data-governance-challenges-that-can-sink-data-operations"&gt;data governance and regulatory compliance risks&lt;/a&gt;. It might also disrupt data loading processes or cause them to fail completely, forcing data teams to scramble to figure out what went wrong.&lt;/p&gt;
 &lt;p&gt;As a result, instances of missing or incomplete data raise complicated data preparation questions. Do they represent substantive data errors? If so, can valid data be inserted? If it can't be, should affected fields be deleted or kept but flagged to show users there are issues with the data?&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;How to overcome this challenge&lt;br&gt;&lt;/b&gt;Effective data profiling identifies missing or incomplete data. Decide what to do about it based on planned use cases and the significance of the data errors. Optimally, data teams or end users should then use a data preparation tool to implement the error-handling measures.&lt;/p&gt;
 &lt;h3&gt;3. Invalid data values&lt;/h3&gt;
 &lt;p&gt;Invalid values are another common data quality issue. They include misspellings, transposed digits, unnecessary characters, duplicate entries and outliers, such as ages, dates and numbers that aren't within a reasonable range. These errors can occur even in enterprise applications with built-in data validation features and end up in analytics and AI data sets.&lt;/p&gt;
 &lt;p&gt;A small number of invalid values in a data set might not have a meaningful impact on applications, but more numerous errors can lead to faulty data analysis results. Cleaning them up should be a priority during data preparation.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;How to overcome this challenge&lt;br&gt;&lt;/b&gt;Finding and fixing invalid data is similar to handling missing values: Profile the data, decide what to do about errors and implement automated functions to address them. Data profiling should also be done on an ongoing basis to identify new issues as data is updated. Perfection is unlikely -- some data errors inevitably slip through. But minimizing them will prevent bad analytics-driven business decisions.&lt;/p&gt;
 &lt;h3&gt;4. Name and address standardization&lt;/h3&gt;
 &lt;p&gt;Inconsistencies in the names, addresses and contact information of consumers and businesses also complicate data preparation. These are legitimate data variations in different systems, not misspellings or missing values. But if not standardized, they can prevent analytics users and AI tools from getting a complete view of customers, suppliers and other business partners.&lt;/p&gt;
 &lt;p&gt;The following are common examples of such inconsistencies:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;A shortened first name or nickname versus a person's full name, such as Fred in one data field and Frederick in another.&lt;/li&gt; 
  &lt;li&gt;Middle initial, full middle name or neither.&lt;/li&gt; 
  &lt;li&gt;Acronyms vs. full business names, such as BMW and Bayerische Motoren Werke.&lt;/li&gt; 
  &lt;li&gt;Companies listed both with and without Inc., Co., Corp., LLC and other business suffixes.&lt;/li&gt; 
  &lt;li&gt;Spelled-out vs. abbreviated address data, such as Boulevard and Blvd. or New York and NY.&lt;/li&gt; 
  &lt;li&gt;Different phone numbers and email addresses for the same entity.&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;p&gt;&lt;b&gt;How to overcome this challenge&lt;br&gt;&lt;/b&gt;Identify inconsistencies through data profiling, then use the standardization features built into a data preparation tool. Alternatively, data teams can create customized standardization processes with a data prep tool's string-handling functionality or use software from a vendor that specializes in name and address standardization.&lt;/p&gt;
 &lt;h3&gt;5. Inconsistent data across enterprise systems&lt;/h3&gt;
 &lt;p&gt;Organizations also encounter inconsistencies when combining data from systems in multiple departments or business units. The data might be correct in each source system, but differences in data formats and entries create problems for analytics and AI applications. It's a pervasive data preparation challenge, especially in large enterprises.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;How to overcome this challenge&lt;br&gt;&lt;/b&gt;When a data attribute, such as an ID field, has different values across source systems, data conversion or cross-reference mapping procedures provide a relatively easy fix. However, if different business rules or data definitions lead to inconsistencies, more complex&amp;nbsp;data transformations are required.&lt;/p&gt;
 &lt;h3&gt;6. Data enrichment issues&lt;/h3&gt;
 &lt;p&gt;Data enrichment helps create the required business context for effective analytics and AI uses. The following are examples of enrichment measures implemented when preparing data:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Augmenting data with entries from other internal or external sources.&lt;/li&gt; 
  &lt;li&gt;Deriving additional data attributes from the existing ones in a data set.&lt;/li&gt; 
  &lt;li&gt;Calculating business metrics and KPIs based on the data.&lt;/li&gt; 
  &lt;li&gt;Organizing data into different structures for planned applications.&lt;/li&gt; 
  &lt;li&gt;Adding tags, labels and metadata to help users understand the data.&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;p&gt;But enriching data isn't easy. Deciding what needs to be done is complicated, and enrichment work can be time-consuming.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;How to overcome this challenge&lt;br&gt;&lt;/b&gt;Data enrichment requires a strong understanding of business needs and goals for the planned applications. Work closely with business executives and users to develop enrichment plans, and allot sufficient resources to the process to meet application delivery schedules.&lt;/p&gt;
 &lt;h3&gt;7. Sustaining and scaling data preparation processes&lt;/h3&gt;
 &lt;p&gt;While data teams and end users sometimes prepare data on an ad hoc basis, data preparation work often becomes a recurring process. Its scope also expands as analytics and AI applications grow and become more widespread -- and valuable -- in enterprises. But organizations often struggle to sustain and scale their data preparation initiatives.&lt;/p&gt;
 &lt;p&gt;Insufficient resources and skills are a problem in some cases. Using custom-coded data preparation methods is, too. If there's no documentation of a custom-coded process, its creator might be the only person who understands how it works, which makes it hard to continue the process if they leave. Also, when modifications to a process are needed, bolting on new code makes maintaining it even more difficult.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;How to overcome this challenge&lt;br&gt;&lt;/b&gt;Ensure that data preparation programs have the required resources and that data teams and end users are properly trained. Using data preparation tools also helps avoid the traps of custom coding. They automatically document processes and &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Data-lineage-documentation-imperative-to-data-quality"&gt;track data lineage and use&lt;/a&gt;, while also providing AI capabilities, collaboration features and connectors to various data sources.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Editor's note:&lt;/b&gt; &lt;i&gt;This article was originally published in 2022. TechTarget editors updated it in March 2026 for timeliness and to add new information.&lt;/i&gt;&lt;/p&gt;
 &lt;p&gt;&lt;i&gt;Rick Sherman, who died in January 2023, was founder and managing partner of Athena Solutions, a BI, data warehousing and data management consulting firm. He had more than 40 years of professional experience in those fields.&lt;/i&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>Data preparation is a crucial but complex part of analytics and AI applications. Don't let these seven common challenges send your data prep processes off track.</description>
            <image>https://cdn.ttgtmedia.com/rms/onlineimages/storage_g539954410.jpg</image>
            <link>https://www.techtarget.com/searchbusinessanalytics/feature/Top-data-preparation-challenges-and-how-to-overcome-them</link>
            <pubDate>Tue, 31 Mar 2026 11:00:00 GMT</pubDate>
            <title>Top data preparation challenges and how to overcome them</title>
        </item>
        <item>
            <body>&lt;p&gt;Enterprises are finding the data infrastructure setups that served them well in the past cannot keep up with today's AI reality.&lt;/p&gt; 
&lt;p&gt;A shift from traditional data architectures to a modern data stack is accelerating thanks to an avalanche of AI initiatives -- and a &lt;a href="https://www.techtarget.com/searchenterpriseai/feature/AI-deployments-gone-wrong-The-fallout-and-lessons-learned"&gt;lack of trust in the data&lt;/a&gt; feeding AI systems. Survey results highlight the problems. Deloitte's 2026 "State of AI in the Enterprise" global survey found that while the number of senior IT and business executives who feel prepared for AI adoption strategically rose to 42 percent from 39 percent the previous year, confidence in their organization's technology infrastructure and data management capabilities declined from 47 percent to 43 percent and from 43 percent to 40 percent, respectively. A 2025 IDC study reported that 84 percent of companies have outdated storage that is not optimal for demanding AI workloads.&lt;/p&gt; 
&lt;p&gt;For enterprise data leaders, it's increasingly a priority to update aging data infrastructure so AI can be deployed with confidence while also modernizing governance and day‑to‑day data management practices that keep AI models reliable and automated decisions defensible.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="From big data complexity to streamlined AI-ready infrastructure"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;From big data complexity to streamlined AI-ready infrastructure&lt;/h2&gt;
 &lt;p&gt;The enterprise data stack is evolving out of necessity. To compete in the AI-first economy, organizations are moving &lt;a href="https://www.techtarget.com/searchdatamanagement/opinion/2026-will-be-the-year-data-becomes-truly-intelligent"&gt;toward data as a product&lt;/a&gt;. This shift replaces brittle, manual workflows with a governed platform designed for scalability, safety and reuse. Under this modern data stack model, IT and data teams provide a secure, shared foundation, while business units maintain ownership of the application outcomes.&lt;/p&gt;
 &lt;p&gt;At each stage of this multilayered approach, data is refined and validated until it is transformed from its raw state into a reusable asset. As organizations roll out autonomous AI agents, this level of granular control over data and &lt;a href="https://www.techtarget.com/searchdatamanagement/post/Key-requirements-for-data-and-analytics-governance-platforms"&gt;comprehensive governance&lt;/a&gt; is a prerequisite for safe, reliable AI applications at scale.&lt;/p&gt;
 &lt;p&gt;Lists of modern data stack layers aren't standardized, and terminology often differs by the source. However, these are its core elements.&lt;/p&gt;
 &lt;h3&gt;1. Ingestion layer&lt;/h3&gt;
 &lt;p&gt;The first layer covers &lt;a href="https://www.techtarget.com/searchbusinessanalytics/tip/6-essential-big-data-best-practices-for-businesses"&gt;data collection&lt;/a&gt; and contains the necessary base infrastructure, including compute resources, networking, cloud services and security controls. In traditional data frameworks, this was largely an IT concern, but it is now a strategic design decision upon which the business goals of data-driven applications rest. It's no longer a choice between on-premises and cloud deployments. Instead, data leaders are designing tailored hybrid infrastructures to distribute processing across on-premises systems for data sovereignty, edge locations for real-time AI performance and cloud environments for scalable compute.&lt;/p&gt;
 &lt;p&gt;Teams can use push or pull methods to ingest data from a wide range of internal and external data sources, such as cloud applications and streaming services. In the modern data stack, there is more of a vetting process. Just because vast amounts of data can be ingested into the infrastructure doesn't mean all of it should be. The modern approach also applies a higher bar for data quality, lineage and provenance. The biggest risk in this stage is fragmentation. If data sources remain disconnected, then teams must manually integrate and clean data and redo engineering work, which slows business processes. &amp;nbsp;&lt;/p&gt;
 &lt;h3&gt;2. Storage layer&lt;/h3&gt;
 &lt;p&gt;In traditional data infrastructure, this layer is often a chaotic catch-all. Companies put their ingested raw data in multiple, disconnected databases, which results in conflicting versions of the truth. This legacy approach makes ensuring AI reliability nearly impossible because there is no single, governed source of information. Data warehouses emerged first to consolidate structured data for BI and fast querying. Later, organizations used data lakes to store unprocessed data to support analytics and AI work. However, operating both a data warehouse and data lake creates redundancies with separate systems for storing and managing different data, which adds to governance and security overhead.&lt;/p&gt;
 &lt;p&gt;&lt;a href="https://www.techtarget.com/searchdatamanagement/feature/The-differences-between-a-data-warehouse-vs-data-mart"&gt;To avoid these data silos&lt;/a&gt; in the modern data stack, organizations are now moving to data lakehouses, which combine the cost efficiency of data lakes with the performance of warehouses. The lakehouse architecture enables unified governance by building a metadata layer that oversees both raw and processed data. Also, by using open table formats to build an organization-wide system of record, companies create a consistent foundation for AI model development. This method improves data processing by reducing the need for unnecessary copies of data and manual engineering.&lt;/p&gt;
 &lt;h3&gt;3. Processing layer&lt;/h3&gt;
 &lt;p&gt;This layer turns the raw data into workable assets, ready to be analyzed or fed into AI models. Processing involves preparing both batch data sets at rest and streaming data in motion for downstream analytics and AI use. This data transformation and curation process includes cleansing, standardizing, enriching, filtering, joining and aggregating the data.&lt;/p&gt;
 &lt;p&gt;In the modern data stack, this layer scales beyond the traditional nightly data update cycle designed for BI dashboard environments. The processing layer must handle real-time updates, &lt;a href="https://www.techtarget.com/searchenterpriseai/definition/multimodal-AI"&gt;multimodal&lt;/a&gt; inputs and automated lineage capture that documents every transformation. This ensures the data's journey from raw to refined is traceable and reduces the risk that AI models will &lt;a href="https://www.techtarget.com/searcherp/podcast/Can-industry-process-models-fix-the-agentic-AI-data-problem"&gt;produce hallucinations and other errors&lt;/a&gt;. Stream processing enables automated alerts and recommendations to be surfaced as quickly as possible so end users and autonomous agents can take immediate actions.&lt;/p&gt;
 &lt;p&gt;Data leaders should ensure their updated infrastructure can handle this additional work without requiring a patchwork of tools and handoffs, which could create governance gaps.&lt;/p&gt;
 &lt;h3&gt;4. Management and distribution layer&lt;/h3&gt;
 &lt;p&gt;In this layer, the processed data is organized so it is fit for purpose. Built-in features work together not just to make the data available but also to ensure it can be governed and discovered. The work here includes data cataloging, lineage visibility, governance policy enforcement and facilitation of data discovery by downstream users.&lt;/p&gt;
 &lt;p&gt;This is the most critical layer and often determines whether the entire modern data stack succeeds or fails. Ultimately, how most businesses operate today depends on data trustworthiness. Gartner predicts that 50 percent of organizations will use a &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/5-benefits-of-building-a-strong-data-governance-strategy"&gt;zero-trust model for data governance&lt;/a&gt; by 2028 due to increasing AI adoption. With the growth of AI-generated data, automated data verification and active metadata management in this layer are essential pieces of the zero-trust governance approach.&lt;/p&gt;
 &lt;p&gt;This layer tends to focus on either data mesh or data fabric architectures, each designed to make it easier for users to locate and share data without added complications. A data mesh is built on distributed domain ownership, where different departments are responsible for their own data under a federated governance structure, while a data fabric uses metadata and automated integration capabilities to join divided data assets and make it easier to reuse them.&lt;/p&gt;
 &lt;h3&gt;5. Context and semantic layer&lt;/h3&gt;
 &lt;p&gt;This is the layer where business logic is applied to both refined and raw data, giving it meaning. This context helps end users, AI systems and automation technologies understand how data should be interpreted across the organization.&lt;/p&gt;
 &lt;p&gt;Shared definitions, knowledge graphs, metrics and other structures provide semantic consistency. Connecting context and semantics to data lineage and access policies reduces decision-making time for users and AI tools alike by removing the need to question whether data is relevant to applications. &amp;nbsp;&lt;/p&gt;
 &lt;h3&gt;6. Integrity and quality layer&lt;/h3&gt;
 &lt;p&gt;This layer maintains the fidelity of data as it moves through the stack. It combines data observability, data stewardship, data quality checks and privacy controls to &lt;a href="https://www.techtarget.com/searchdatamanagement/opinion/Data-contracts-help-build-trustworthy-data-products-for-AI"&gt;ensure data is accurate&lt;/a&gt;, consistent, documented and protected for effective decision-making.&lt;/p&gt;
 &lt;p&gt;This arrangement provides structure to the stack to prevent unreliable data feeds and data silos. Data quality rules identify missing values, data duplication and freshness issues. Master data management practices create common records for business entities, such as customers and products, to maintain consistency across systems. Data stewards apply governance and security policies that &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Data-governance-challenges-that-can-sink-data-operations"&gt;dictate who gets access to data and when&lt;/a&gt;.&lt;/p&gt;
 &lt;h3&gt;7. Consumption layer&lt;/h3&gt;
 &lt;p&gt;This is the top of the stack, the culmination of all the architectural choices designed to produce refined, trusted data and get it to the right users and systems at the right time. &amp;nbsp;&lt;/p&gt;
 &lt;p&gt;Traditionally, consumption meant dashboards, reports and analytics tools, but it now includes embedded analytics, machine learning applications, and agentic AI or semi-autonomous workflows. Rather than simply adding AI to old processes, data leaders are redesigning this layer so &lt;a href="https://www.techtarget.com/searchsoftwarequality/tip/How-effective-is-your-AI-agent-benchmarks-to-consider"&gt;agents and people can work collaboratively&lt;/a&gt; with clear decision-making boundaries, ensuring IT provides the platform while business units determine results.&lt;/p&gt;
 &lt;div class="youtube-iframe-container"&gt;
  &lt;iframe id="ytplayer-0" src="https://www.youtube.com/embed/7FufIRExfpo?autoplay=0&amp;amp;modestbranding=1&amp;amp;rel=0&amp;amp;widget_referrer=null&amp;amp;enablejsapi=1&amp;amp;origin=https://www.techtarget.com" type="text/html" height="360" width="640" frameborder="0"&gt;&lt;/iframe&gt;
 &lt;/div&gt;
&lt;/section&gt;                            
&lt;section class="section main-article-chapter" data-menu-title="What matters most when reassessing the data stack"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;What matters most when reassessing the data stack&lt;/h2&gt;
 &lt;p&gt;When it's time to update how your organization processes data and data platform vendors come calling, prepare product evaluation questions to meet your specific needs rather than getting lost in talks about performance and feature checklists.&lt;/p&gt;
 &lt;p&gt;AI initiatives introduce a new set of requirements beyond the capabilities of existing data architectures. Today, the priorities include avoiding data duplication, improved data portability, strong lineage and consistency across departments and clouds.&lt;/p&gt;
 &lt;p&gt;Tailor these modern data stack platform requirements for your organization, but these are some questions to ask:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Does the platform provide a unified semantic layer and active metadata to ensure consistent logic across AI agents and BI applications?&lt;/li&gt; 
  &lt;li&gt;Does the platform support hybrid cloud and multi-cloud deployments by design for seamless workload migration based on cost, performance or data sovereignty requirements?&lt;/li&gt; 
  &lt;li&gt;Does it have policy-as-code &lt;a target="_blank" href="https://www.cncf.io/blog/2025/07/29/introduction-to-policy-as-code/" rel="noopener"&gt;capabilities&lt;/a&gt; to standardize data governance, privacy and quality across data assets, and AI models and agents?&lt;/li&gt; 
  &lt;li&gt;What are the platform's capabilities related to open table formats, APIs and portable pipelines to avoid extensive work when moving data and workloads?&lt;/li&gt; 
  &lt;li&gt;What is the status of agentic AI governance, and what are the plans to close any oversight gaps?&lt;/li&gt; 
  &lt;li&gt;Is there a single management interface for data stewards to monitor policy enforcement and issue resolution?&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="What's coming next for the modern data stack?"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;What's coming next for the modern data stack?&lt;/h2&gt;
 &lt;p&gt;All signals from leading analyst firms indicate the next evolution of the data stack will refine context awareness, tighten governance and integrate more closely with business workflows and agentic AI systems. These trends are linked: as companies increasingly deploy agents, they need richer context and stronger data controls. Deloitte's 2026 AI survey &lt;a target="_blank" href="https://www.deloitte.com/us/en/what-we-do/capabilities/applied-artificial-intelligence/content/state-of-ai-in-the-enterprise.html" rel="noopener"&gt;found&lt;/a&gt; that while 74 percent of companies plan to deploy agentic AI within two years, only 21 percent have a governance model for them now.&lt;/p&gt;
 &lt;p&gt;Vendors are converging the stack, joining layers, &lt;a href="https://www.techtarget.com/searchdatamanagement/news/366631576/New-consortium-to-aid-AI-by-standardizing-semantic-modeling"&gt;improving semantic structure&lt;/a&gt; and embedding oversight. They are moving toward a unified, governed data lakehouse to reduce redundant copies and data movement across silos, cutting costs and security risks. This architecture supports the federated, shared ownership model in which business leaders set standards and quality expectations, while IT manages the data lakehouse and enforces policies to keep data and AI aligned at scale.&lt;/p&gt;
 &lt;p&gt;For organizations reassessing their existing stack architecture, take a modular approach. Avoid overbuying and focus on the immediate needs for data context and trust. This provides flexibility to get AI and analytics work done today rather than a rigid, expensive redesign that might be obsolete in a few years.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Editor's note&lt;/b&gt;&lt;i&gt;: TechTarget editors updated this article, originally published in 2023 and written by &lt;a href="https://www.techtarget.com/contributor/Jeff-McCormick"&gt;Jeff McCormick&lt;/a&gt;, in March 2026 to add new information and improve timeliness.&lt;/i&gt;&lt;/p&gt;
 &lt;p&gt;&lt;i&gt;Tom Walat is an editor and reporter for TechTarget, where he covers data technologies.&lt;/i&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>Data infrastructure and practices need an upgrade for the AI era. A modern data stack takes a layered approach that aligns teams and delivers governed, trusted data.</description>
            <image>https://cdn.ttgtmedia.com/rms/onlineimages/container_g488602622.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/tip/Assemble-the-layers-of-big-data-stack-architecture</link>
            <pubDate>Fri, 20 Mar 2026 10:24:00 GMT</pubDate>
            <title>Understanding the layers of the AI‑ready modern data stack</title>
        </item>
        <item>
            <body>&lt;p&gt;Smart organizations use big data to better understand their customers, identify market trends and improve business operations, thereby boosting financial performance and gaining a competitive advantage over rivals. However, investing in big data technologies and applications without a strategic plan is a recipe for wasting time, money and resources.&lt;/p&gt; 
&lt;p&gt;Without a well-defined strategy for &lt;a href="https://www.techtarget.com/searchdatamanagement/The-ultimate-guide-to-big-data-for-businesses"&gt;managing and using big data assets&lt;/a&gt;, a company might end up with separate, uncoordinated initiatives. There's a risk of duplicate or conflicting analytics and AI projects, as well as ones that aren't aligned with strategic business objectives.&lt;/p&gt; 
&lt;p&gt;Developing a comprehensive strategy to underpin &lt;a href="https://www.techtarget.com/searchbusinessanalytics/feature/8-big-data-use-cases-for-businesses-and-industry-examples"&gt;big data applications&lt;/a&gt; is easier said than done, but the guidance and steps outlined below will help data leaders manage the process.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="What a big data strategy includes"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;What a big data strategy includes&lt;/h2&gt;
 &lt;p&gt;An effective big data strategy maps out how the data will be used to support business processes and decision-making. It defines specific business goals for big data applications and sets guidelines for using data to ensure compliance with privacy and regulatory requirements. To align big data initiatives with business needs and objectives, business leaders must be involved in developing the strategy from start to finish.&lt;/p&gt;
 &lt;p&gt;The strategy also specifies procedures for managing the big data environment. That includes details on how data management and analytics teams will &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/10-big-data-challenges-and-how-to-address-them"&gt;address various big data challenges&lt;/a&gt;, such as the following:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Collecting and processing a combination of unstructured, semistructured and structured data from both internal and external sources.&lt;/li&gt; 
  &lt;li&gt;&lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Establish-big-data-integration-techniques-and-best-practices"&gt;Integrating different data sets&lt;/a&gt; to give end users a comprehensive view of relevant data.&lt;/li&gt; 
  &lt;li&gt;Identifying and &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Data-quality-for-big-data-Why-its-a-must-and-how-to-improve-it"&gt;fixing data quality problems&lt;/a&gt; to ensure data is accurate, consistent and trustworthy.&lt;/li&gt; 
  &lt;li&gt;Controlling storage and associated data management costs.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="Building a big data strategy"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Building a big data strategy&lt;/h2&gt;
 &lt;p&gt;Here are four steps to take when formulating a big data strategy for an organization.&lt;/p&gt;
 &lt;h3&gt;1. Define business goals and objectives for big data applications&lt;/h3&gt;
 &lt;p&gt;Start by defining the business objectives that the strategy aims to achieve. Businesses aren't the same, so there's no one-size-fits-all answer here. Align the strategy with corporate business objectives, critical KPIs and key business problems the company needs to address.&lt;/p&gt;
 &lt;p&gt;Input from senior executives and business managers on business goals and needs ensures that the strategy supports their priorities -- and that the organization will adopt it. Also involve data scientists and analysts who work with the business on analytics initiatives, as well as members of the data management team.&lt;/p&gt;
 &lt;h3&gt;2. Identify relevant data sources and assess data readiness&lt;/h3&gt;
 &lt;p&gt;The next step is identifying useful data sources to incorporate into the strategy and assessing the readiness of their data assets. As part of the assessment, document data formats, profile data, measure quality levels and evaluate data integration and transformation requirements.&lt;/p&gt;
 &lt;p&gt;Map data sources to the strategy's business objectives and gauge data readiness accordingly. For example, if improving CX is a business objective, the readiness assessment should cover any data assets related to customer touchpoints.&lt;/p&gt;
 &lt;h3&gt;3. Identify and prioritize big data use cases&lt;/h3&gt;
 &lt;p&gt;Think big on use cases, but start small when developing plans for big data applications. Be realistic about how much the data management and analytics teams can handle at once. Upfront analytics can help identify applicable -- and achievable -- use cases by uncovering patterns, correlations and other useful data insights.&lt;/p&gt;
 &lt;p&gt;Prioritize use cases based on factors such as their &lt;a href="https://www.techtarget.com/searchbusinessanalytics/feature/6-big-data-benefits-for-businesses"&gt;potential business benefits&lt;/a&gt; and the required budget and resources. Depending on the number of departments and business units involved, this process can be complex. Work with the various stakeholders to create a plan, then document which use cases will be pursued so everyone is aware of the prioritized list.&lt;/p&gt;
 &lt;h3&gt;4. Create a roadmap for big data projects&lt;/h3&gt;
 &lt;p&gt;Plotting a roadmap for big data applications is often the most time-consuming step when building a big data strategy. Even after the roadmap is completed, it isn't set in stone. It will likely evolve over time as business objectives, priorities and opportunities change.&lt;/p&gt;
 &lt;p&gt;As part of the roadmap, identify gaps &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Building-a-big-data-architecture-Core-components-best-practices"&gt;&lt;/a&gt;in data technologies, processes and skill sets that could affect the success of planned applications. The gap analysis will inform investments in the &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Building-a-big-data-architecture-Core-components-best-practices"&gt;big data architecture&lt;/a&gt; and the internal resources needed to support the applications. It might also prompt a review of the prioritized use cases to assess whether any changes are needed due to existing gaps that can't be filled immediately.&lt;/p&gt;
 &lt;figure class="main-article-image full-col" data-img-fullsize="https://www.techtarget.com/rms/onlineimages/datamanagement-4_steps_to_building_a%20big_data_strategy-f.png"&gt;
  &lt;img data-src="https://www.techtarget.com/rms/onlineimages/datamanagement-4_steps_to_building_a%20big_data_strategy-f_mobile.png" class="lazy" data-srcset="https://www.techtarget.com/rms/onlineimages/datamanagement-4_steps_to_building_a%20big_data_strategy-f_mobile.png 960w,https://www.techtarget.com/rms/onlineimages/datamanagement-4_steps_to_building_a%20big_data_strategy-f.png 1280w" alt="Graphic containing text that describes four key steps for building a big data strategy." height="301" width="559"&gt;
  &lt;figcaption&gt;
   &lt;i class="icon pictures" data-icon="z"&gt;&lt;/i&gt;Organizations should follow these steps to create a big data strategy.
  &lt;/figcaption&gt;
  &lt;div class="main-article-image-enlarge"&gt;
   &lt;i class="icon" data-icon="w"&gt;&lt;/i&gt;
  &lt;/div&gt;
 &lt;/figure&gt;
&lt;/section&gt;               
&lt;section class="section main-article-chapter" data-menu-title="Be flexible when implementing the strategy"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Be flexible when implementing the strategy&lt;/h2&gt;
 &lt;p&gt;Flexibility is often the most important principle to adopt when implementing a big data strategy. Business needs, data and &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/15-big-data-tools-and-technologies-to-know-about"&gt;available tools and technologies&lt;/a&gt; aren't static, so strategy development isn't a one-and-done exercise.&lt;/p&gt;
 &lt;p&gt;Data leaders must be prepared to quickly adjust budgets, technologies, staffing and data management and analytics processes in response to changing circumstances. For example, IT infrastructure changes might be necessary to ensure end users can access critical data from new sources. Increasing deployments of enterprise AI applications also create new data demands that must be &lt;a href="https://www.techtarget.com/searchenterpriseai/tip/How-do-big-data-and-AI-work-together"&gt;factored into big data strategies&lt;/a&gt;.&lt;/p&gt;
 &lt;p&gt;Similarly, required roles and skill sets might change over time. As a result, building strong teams to support big data initiatives typically relies on a combination of external hiring and retraining or upskilling of current employees. The balance between those approaches will likely fluctuate depending on particular staffing needs.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Editor's note:&lt;/b&gt; &lt;i&gt;TechTarget editors updated this article in March 2026 for timeliness and to add new information.&lt;/i&gt;&lt;/p&gt;
 &lt;p&gt;&lt;i&gt;Kathleen Walch is director of AI engagement and community at Project Management Institute. She previously was co-founder and managing partner of Cognilytica, a technology research and training firm acquired by PMI in 2024.&lt;/i&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>Big data initiatives won't deliver business benefits without a comprehensive strategy to guide data management and analytics work. Here's how to build one.</description>
            <image>https://cdn.ttgtmedia.com/visuals/searchContentManagement/collaboration_technology/contentmanagement_article_003.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/feature/How-to-build-an-enterprise-big-data-strategy-in-4-steps</link>
            <pubDate>Thu, 19 Mar 2026 16:34:00 GMT</pubDate>
            <title>How to build an effective big data strategy</title>
        </item>
        <item>
            <body>&lt;p&gt;Business leaders generally know what data they have, where it's stored and who can access it. What's less visible, but increasingly important, is how that data changed over time.&lt;/p&gt; 
&lt;p&gt;&lt;a href="https://www.techtarget.com/searchdatamanagement/tip/How-data-lineage-tools-boost-data-governance-policies"&gt;Data lineage&lt;/a&gt; provides visibility into a data asset's complete lifecycle, including its origin, transformations and access history. This context strengthens risk mitigation, compliance readiness and data quality efforts. It's like having every set of blueprints for a house, from the initial build through every renovation; there are no secrets or surprises. When leaders can trace how business decisions were made and how their data changed over time, they gain confidence in the tools, people and information involved in determining what should change moving forward.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="Why businesses need data lineage"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Why businesses need data lineage&lt;/h2&gt;
 &lt;p&gt;Many organizations focus on present and future data use, but the past matters just as much. Without visibility into how data is created, changed and shared, teams struggle to maintain data quality, transparency and reliability -- all of which limit effective decision-making.&lt;/p&gt;
 &lt;p&gt;Effective data lineage supports the following:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;&lt;b&gt;Accountability.&lt;/b&gt; Clear lineage shows who accessed or changed data, helping identify who made improvements and who introduced risks. For example, a data engineer who&lt;a href="https://www.techtarget.com/searchdatamanagement/tip/6-dimensions-of-data-quality-boost-data-performance"&gt; improved data quality&lt;/a&gt; versus someone who made malicious changes.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Data quality.&lt;/b&gt; Tracing transformations identifies &lt;a href="https://www.informationweek.com/data-management/11-irritating-data-quality-issues" target="_blank" rel="noopener"&gt;the stage where data quality dropped&lt;/a&gt;, such as deletions, additions, deduplications or merges, allowing teams to fix issues at the source.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Reduced technical debt.&lt;/b&gt; Understanding data's original purpose highlights obsolete infrastructure or unused assets that can be safely deleted or archived.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Compliance.&lt;/b&gt; Regulatory frameworks increasingly expect traceable and documented data flows. For example, the &lt;a href="https://www.techtarget.com/searchenterpriseai/opinion/Everything-you-need-to-know-about-the-new-EU-AI-Act"&gt;EU AI Act&lt;/a&gt; requires organizations to trace the provenance of an AI model's training data. While more traditional regulations, such as GDPR and HIPAA, do not explicitly require data lineage, tracking the origin and lifecycle of data is critical for meeting their privacy and auditability standards.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="The risks of poor data lineage"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;The risks of poor data lineage&lt;/h2&gt;
 &lt;p&gt;When businesses fail to maintain adequate visibility into data lineage, they expose themselves to financial and reputational risk. Without data lineage, organizations struggle with:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;&lt;b&gt;Limited decision-making context.&lt;/b&gt; Leaders cannot validate insights when they lack visibility of the data lifecycle. This gap can disrupt &lt;a href="https://www.techtarget.com/searchbusinessanalytics/tip/Key-steps-form-a-data-driven-decision-making-framework"&gt;effective decision-making.&lt;/a&gt; It's hard to have confidence in an analytics report without the historical perspective to know whether metrics were removed, filtered or dropped along the way, which can obscure growth opportunities.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Erosion of stakeholder trust.&lt;/b&gt; Unclear data origins undermine stakeholder confidence in business processes or decisions that depend on data. Earning back that trust can be a lengthy, costly process.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Compliance exposure.&lt;/b&gt; Many regulations require traceable data handling. Businesses without these records can face audits, sanctions or lawsuits.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Lower data quality.&lt;/b&gt; Data quality typically suffers when businesses lose visibility into data's origins and evolution, making it difficult to find and fix errors.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Reduced IT efficiency.&lt;/b&gt; Incomplete lineage makes it harder to &lt;a href="https://www.techtarget.com/searchcio/feature/Key-technical-debt-reduction-strategies"&gt;manage technical debt&lt;/a&gt; effectively, reducing IT and engineering teams' ability to focus on more innovative work.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="The AI imperative for data lineage"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;The AI imperative for data lineage&lt;/h2&gt;
 &lt;p&gt;While maintaining data lineage is important for every data asset, growing adoption of &lt;a href="https://www.techtarget.com/searchenterpriseai/tip/Agentic-AI-vs-generative-AI-Whats-the-difference"&gt;generative and agentic AI technology&lt;/a&gt; has made it even more critical. These technologies depend on AI and machine learning models trained on vast amounts of data. Without knowing data's origins or how it evolved, troubleshooting and optimizing AI models becomes significantly more difficult.&lt;/p&gt;
 &lt;p&gt;&lt;a href="https://www.informationweek.com/machine-learning-ai/ai-hallucinations-can-prove-costly" target="_blank" rel="noopener"&gt;Hallucinations are a big risk&lt;/a&gt;. When a model frequently produces unreliable outputs, then lineage helps determine whether the issue is related to insufficient, missing or altered data, or a flaw in the model design.&lt;/p&gt;
 &lt;p&gt;Beyond working with training data, data lineage also affects &lt;a href="https://www.nojitter.com/ai-voice/managing-ai-tools-requires-more-data-transparency" target="_blank" rel="noopener"&gt;AI model deployment and management&lt;/a&gt;. It provides visibility into prompt libraries that organizations use to guide how users interact with models. Data lineage is useful for complex, multi-step agentic workflows because it tracks how data evolves as it interacts with AI agents.&lt;/p&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="How to achieve effective data lineage"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;How to achieve effective data lineage&lt;/h2&gt;
 &lt;p&gt;Creating effective data lineage involves the following core processes:&lt;/p&gt;
 &lt;ol class="default-list"&gt; 
  &lt;li&gt;&lt;b&gt;Identify valuable assets.&lt;/b&gt; Not every asset requires lineage, such as short-lived data. Focus on high-value data assets.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Set lineage depth.&lt;/b&gt; Determine the amount of detail required. Some data assets require more extensive lineage than others.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Record data events.&lt;/b&gt; After establishing lineage goals, organizations should record all data-related events that fall within the scope of their plans, including data creation, transformation and changes to access policies.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Validate lineage. &lt;/b&gt;Periodically audit lineage records to detect and correct visibility gaps or enforcement oversights.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Adapt over time.&lt;/b&gt; Revisit lineage needs and strategy regularly and update practices as needed.&lt;/li&gt; 
 &lt;/ol&gt;
 &lt;p&gt;Automated lineage tools help track data history, but they require human oversight to ensure accuracy. Organizations often assign data stewards to this role, and data engineers and &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Data-governance-roles-and-responsibilities-Whats-needed"&gt;data governance teams&lt;/a&gt; typically share responsibility to maintain lineage quality.&lt;/p&gt;
 &lt;p&gt;&lt;em&gt;Chris Tozzi is a freelance writer, research adviser, and professor of IT and society who has previously worked as a journalist and Linux systems administrator.&lt;/em&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>Data lineage is critical to enterprise success. Tracing a data asset's journey through pipelines helps improve data quality, speeds up corrections and builds trust in analytics.</description>
            <image>https://cdn.ttgtmedia.com/visuals/searchDataManagement/content_management/datamanagement_article_003.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/tip/Data-lineage-documentation-imperative-to-data-quality</link>
            <pubDate>Wed, 18 Mar 2026 10:00:00 GMT</pubDate>
            <title>Data lineage documentation matters for enterprise reliability</title>
        </item>
        <item>
            <body>&lt;p&gt;Open source databases have steadily gained ground on proprietary systems, driven by the rise of Linux, cloud computing and NoSQL technologies. Open source databases and source-available alternatives provide organizations with options beyond proprietary platforms.&lt;/p&gt; 
&lt;p&gt;Open source software offers user organizations the promise of source code developed in the open, typically in a community-driven process. The aim is to expand the number of people involved in the development process and not lock users into a specific vendor's technology.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="What are open source databases?"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;What are open source databases?&lt;/h2&gt;
 &lt;p&gt;&lt;a name="_Hlk164960278"&gt;&lt;/a&gt;Open source databases are developed and released under an open source license. While &lt;i&gt;open source&lt;/i&gt; is sometimes used as a marketing term, it has a very specific definition when it comes to software licenses. To qualify as open source, a database must use a license approved by the Open Source Initiative. The OSI determines whether licenses adhere to the Open Source Definition (OSD), which is the &lt;a href="https://opensource.org/osd"&gt;guiding document&lt;/a&gt; for open source licensing.&lt;/p&gt;
 &lt;p&gt;Licensing has become more complicated, though. A growing number of vendors that created open source databases have adopted licenses that largely adhere to the tenets of the OSD but aren't OSI-approved. In many cases, these licenses require cloud providers offering database as a service (DBaaS) implementations to release modified or related source code under the same license. These are typically called source-available licenses. Databases licensed this way are often grouped with fully open source technologies as alternatives to proprietary, closed source databases.&lt;/p&gt;
 &lt;p&gt;The broad category of open source and source-available databases contains &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Evaluating-the-different-types-of-DBMS-products"&gt;various types of database software&lt;/a&gt; that support different applications. That includes SQL-based &lt;a href="https://www.techtarget.com/searchdatamanagement/definition/relational-database"&gt;relational databases&lt;/a&gt;, the most widely used type, and the four primary NoSQL technologies -- key-value stores, document databases, wide-column stores and &lt;a href="https://www.techtarget.com/whatis/definition/graph-database"&gt;graph databases&lt;/a&gt;. Open source versions of special-purpose systems, such as vector databases and time series databases, are also available. In addition, many vendors now offer databases that support more than one data model.&lt;/p&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="Potential benefits of using open source databases"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Potential benefits of using open source databases&lt;/h2&gt;
 &lt;p&gt;Open source databases &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Top-open-source-database-advantages-for-enterprises"&gt;offer many potential benefits&lt;/a&gt;, some of which also apply to source-available technologies. The following are among the primary benefits for user organizations:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;&lt;b&gt;Easy to get started. &lt;/b&gt;A core premise of the open source approach is that the technology is freely available. As a result, users can easily try out and deploy an open source database without first having to pay for it. Vendors do offer paid support as well as closed source versions of databases with additional features in many cases.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Community support and engagement.&lt;/b&gt; Open source or source-available code typically comes with a community of engaged users and contributors who can help new users with the technology. It also enables a degree of participation in the code development process. For example, users can submit bug reports and feature requests and become contributors themselves.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Source code transparency.&lt;/b&gt; When source code is open and can be viewed by anyone, there's a better chance of understanding how a database works and how it can be used effectively to meet business needs.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Flexibility and customization.&lt;/b&gt; With some open source licenses, developers are free to modify the database software to meet specific custom requirements.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Improved security.&lt;/b&gt; Because the source code is open, developers, users and security researchers can thoroughly scrutinize it to identify vulnerabilities. That enables &lt;a href="https://www.techtarget.com/searchsecurity/tip/5-enterprise-patch-management-best-practices"&gt;rapid patching of vulnerabilities&lt;/a&gt; after they're discovered.&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;p&gt;The technologies listed below are among the most prominent open source and source-available databases. TechTarget editors compiled the list based on research into the database market, including Gartner vendor rankings and database management system (&lt;a href="https://www.techtarget.com/searchdatamanagement/definition/database-management-system"&gt;DBMS&lt;/a&gt;) popularity rankings from DB-Engines. However, the list itself is unranked and includes five relational open source databases, three NoSQL ones and four source-available technologies, organized in that order.&lt;/p&gt;
 &lt;p&gt;Each write-up outlines key features, potential use cases, licensing and commercial support options to help organizations choose the right database for their application needs.&lt;/p&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="1. MySQL"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;1. MySQL&lt;/h2&gt;
 &lt;p&gt;MySQL is among the most widely deployed open source databases. It was first released in 1996 as an independent effort led by Michael "Monty" Widenius and two other developers, who co-founded MySQL AB to create the database. The company was acquired in 2008 by Sun Microsystems, which was then bought by Oracle in 2010. MySQL has remained a core part of Oracle's database portfolio ever since while being maintained as open source software.&lt;/p&gt;
 &lt;p&gt;A relational database, MySQL was originally positioned as an online transaction processing (OLTP) system and is still primarily geared to transactional uses, although Oracle's MySQL HeatWave cloud database service now also supports analytics and &lt;a href="https://www.techtarget.com/searchenterpriseai/definition/machine-learning-ML"&gt;machine learning&lt;/a&gt; applications. MySQL gained much of its early popularity as a cornerstone of the LAMP stack of open source technologies -- Linux, Apache, MySQL and PHP, Perl or Python -- that powered the first generation of web development. It continues to be an underlying database on many websites today.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Common use cases: &lt;/b&gt;Like other relational databases, MySQL complies with the &lt;a href="https://www.techtarget.com/searchdatamanagement/definition/ACID"&gt;ACID&lt;/a&gt; properties -- atomicity, consistency, isolation and durability – to ensure data integrity and reliability. As a result, it supports a broad range of applications. For example, MySQL is commonly used as a web application server and to run cloud applications and content management systems.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Licensing: &lt;/b&gt;MySQL is dual-licensed under the GPL version 2 open source license and an Oracle license for organizations looking to distribute the database along with commercial applications.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Source code repository&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; &lt;a target="_blank" href="https://github.com/mysql/mysql-server" rel="noopener"&gt;https://github.com/mysql/mysql-server&lt;/a&gt;&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Commercial support options: &lt;/b&gt;There are numerous commercial implementations of MySQL. Oracle offers multiple options beyond MySQL Heatwave, including Enterprise and Standard editions and an embedded version. MySQL is also available in the cloud as part of the &lt;a href="https://www.techtarget.com/searchaws/definition/Amazon-Relational-Database-Service-RDS"&gt;Amazon Relational Database Service (RDS)&lt;/a&gt; from AWS, as well as Google's Cloud SQL and Microsoft's Azure Database services. Vendors such as Aiven, PlanetScale, and Percona also offer MySQL cloud services.&lt;/p&gt;
&lt;/section&gt;       
&lt;section class="section main-article-chapter" data-menu-title="2. MariaDB"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;2. MariaDB&lt;/h2&gt;
 &lt;p&gt;MariaDB debuted in 2009 as a fork of MySQL that was created by a team also led by Widenius, who left Sun early that year because he was concerned about the direction and development of MySQL. Work on MariaDB started when he was still at Sun, and it was originally designed to be a drop-in replacement for MySQL. But that was only fully true until the 5.5 releases of the two databases. After that, new features not in MySQL were added to MariaDB, which used different numbering on subsequent releases.&lt;/p&gt;
 &lt;p&gt;Even with newer updates, though, it's still relatively easy to migrate from MySQL to MariaDB. The latter's data files are generally binary-compatible with MySQL's, and the database client protocols are also compatible. As a result, users in many cases can simply uninstall MySQL and install MariaDB to change between them. MariaDB PLC, which leads development of the software through the MariaDB Foundation, maintains a list of &lt;a target="_blank" href="https://mariadb.com/kb/en/mariadb-vs-mysql-compatibility/" rel="noopener"&gt;incompatibilities and feature differences&lt;/a&gt; with MySQL.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Common use cases:&lt;/b&gt; MariaDB is commonly used for the same purposes as MySQL, including in web and cloud applications involving both transaction processing and analytics workloads.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Licensing:&lt;/b&gt; The free MariaDB Server software -- referred to by the company as MariaDB Community Server -- is released under the GPLv2 license.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Source code repository&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; &lt;a target="_blank" href="https://github.com/MariaDB/server" rel="noopener"&gt;https://github.com/MariaDB/server&lt;/a&gt;&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Commercial support options&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; MariaDB PLC sells a MariaDB Enterprise Server version of the database with enterprise features such as &lt;a href="https://www.theserverside.com/definition/JSON-Javascript-Object-Notation"&gt;JSON&lt;/a&gt; support and columnar storage. MariaDB Cloud is the company's fully managed DBaaS offering. MariaDB is also available through Amazon RDS, while Microsoft retired Azure Database for MariaDB in September 2025.&lt;/p&gt;
&lt;/section&gt;       
&lt;section class="section main-article-chapter" data-menu-title="3. PostgreSQL"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;3. PostgreSQL&lt;/h2&gt;
 &lt;p&gt;PostgreSQL got its start as Postgres in 1986 at the University of California, Berkeley. The Postgres project was initiated by relational database pioneer Michael Stonebraker, then a professor at the school, as a more advanced alternative to Ingres, a proprietary relational database management system (&lt;a href="https://www.techtarget.com/searchdatamanagement/definition/RDBMS-relational-database-management-system"&gt;RDBMS&lt;/a&gt;) that he also played a lead role in developing. The software became open source in 1995, when a SQL language interpreter was also added, and it was officially renamed PostgreSQL in 1996. Decades later, though, PostgreSQL and Postgres are still used interchangeably by developers, vendors and users to refer to the database.&lt;/p&gt;
 &lt;p&gt;PostgreSQL offers full RDBMS features, including ACID compliance, SQL querying and support for procedural language queries to create stored procedures and triggers in databases. Like MySQL, MariaDB and many other database technologies, it also supports multiversion concurrency control (&lt;a href="https://www.theserverside.com/blog/Coffee-Talk-Java-News-Stories-and-Opinions/What-is-MVCC-How-does-Multiversion-Concurrencty-Control-work"&gt;MVCC&lt;/a&gt;) so data can be read and updated by different users at the same time. In addition, PostgreSQL supports other types of database objects than standard relational tables, and it's described as an object-relational DBMS on the open source project's website.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Common use cases: &lt;/b&gt;PostgreSQL is commonly positioned as an open source alternative to the proprietary Oracle Database. It's widely used to support enterprise applications that require complex transactions and high levels of concurrency, and sometimes in data warehousing operations.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Licensing&lt;/b&gt;: The software is available under the OSI-approved PostgreSQL License.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Source code repository&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; &lt;a target="_blank" href="https://git.postgresql.org/gitweb/?p=postgresql.git;a=summary" rel="noopener"&gt;https://git.postgresql.org/gitweb/?p=postgresql.git;a=summary&lt;/a&gt;&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Commercial support options:&lt;/b&gt; PostgreSQL has a wide range of commercial support and managed cloud offerings. EDB specializes in PostgreSQL and provides both self-managed and DBaaS versions in the cloud. Managed PostgreSQL cloud services are also available from &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Cloud-database-comparison-AWS-Microsoft-Google-and-Oracle"&gt;AWS, Google, Microsoft and Oracle&lt;/a&gt;, as well as vendors such as Aiven, Percona and Instaclustr.&lt;/p&gt;
&lt;/section&gt;       
&lt;section class="section main-article-chapter" data-menu-title="4. Firebird"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;4. Firebird&lt;/h2&gt;
 &lt;p&gt;The Firebird open source relational database's technology roots go back to the early 1980s, when the proprietary InterBase database was created. After InterBase was acquired by multiple vendors, commercial product development ended and the final release was made available under an open source license in 2000. Within a week, the Firebird project was created to continue developing a fork of the technology.&lt;/p&gt;
 &lt;p&gt;Firebird supports ACID-compliant transactions, external user-defined functions and various standard SQL features, and it includes a multi-generational architecture that provides MVCC capabilities. The software has a relatively small footprint and is available in an embedded single-user version, but it can also be used to run multi-terabyte databases with hundreds of concurrent users. It shouldn't be confused with Firestore and Firebase Realtime Database, two commercial NoSQL databases developed by Google.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Common use cases:&lt;/b&gt; Firebird can handle both operational and analytics applications. It's used in various types of enterprise applications, including ERP and CRM systems.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Licensing&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; Firebird is made available under the InterBase Public License (IPL) and the Initial Developer's Public License (IDPL). Both are variants of the Mozilla Public License Version 1.1, which is OSI-approved though now superseded by Version 2.0. The IPL covers the source code from InterBase, while the IDPL applies to added or improved code developed as part of the Firebird project.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Source code repository&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; &lt;a target="_blank" href="https://github.com/FirebirdSQL/firebird" rel="noopener"&gt;https://github.com/FirebirdSQL/firebird&lt;/a&gt;&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Commercial support options: &lt;/b&gt;Firebird is an independent open source project, and the software is free to use, including for commercial purposes. The Firebird project points users to commercial providers for technical support, consulting and training services. Firebird cloud services running on Windows Server 2019 are available for purchase in the AWS, Azure and Google clouds, although support ended on Google Cloud one in August 2024.&lt;/p&gt;
&lt;/section&gt;       
&lt;section class="section main-article-chapter" data-menu-title="5. SQLite"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;5. SQLite&lt;/h2&gt;
 &lt;p&gt;SQLite is a lightweight embedded RDBMS that runs inside applications. It was created in 2000 by computer analyst and programmer D. Richard Hipp while he was working as a government contractor supporting a U.S. Navy project, which needed a database that could run without a database administrator (&lt;a href="https://www.techtarget.com/searchdatamanagement/definition/database-administrator"&gt;DBA&lt;/a&gt;) in environments with minimal resources. Hipp continues to lead development of the software as project architect through Hipp, Wyrick &amp;amp; Company Inc., a software engineering firm commonly known as Hwaci.&lt;/p&gt;
 &lt;p&gt;As an embedded database, SQLite is self-contained, meaning it's fully functional within the application it powers. The software is a library that embeds a full-featured SQL database engine supporting ACID transactions. There are no separate database server processes. Data reads and writes are done directly to ordinary disk files, and a complete SQLite database that includes tables, indices, triggers and views can be contained in a single file.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Common use cases:&lt;/b&gt; SQLite is commonly used in mobile applications, web browsers and IoT devices due to its small footprint and ability to operate without a separate server process.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Licensing:&lt;/b&gt; The SQLite source code is in the public domain and is free to use, modify and distribute for any purpose without a license. Hwaci does sell a warranty of title with a perpetual right-to-use license to organizations that want one for legal reasons.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Source code repository&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; &lt;a target="_blank" href="https://sqlite.org/src/doc/trunk/README.md" rel="noopener"&gt;https://sqlite.org/src/doc/trunk/README.md&lt;/a&gt;&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Commercial support options:&lt;/b&gt; Hwaci provides paid technical support, maintenance and testing services, and it offers a set of proprietary extensions to SQLite that are sold under separate licenses. As with Firebird, SQLite database services are available on AWS, Azure and Google Cloud, though the underlying marketplace images vary by platform.&lt;/p&gt;
&lt;/section&gt;       
&lt;section class="section main-article-chapter" data-menu-title="6. Apache Cassandra"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;6. Apache Cassandra&lt;/h2&gt;
 &lt;p&gt;The Cassandra wide-column store traces its roots back to 2007, when Facebook developed it to support a new inbox search feature. The NoSQL database was converted to open source in 2008 and became part of the Apache Software Foundation in 2009, initially as an incubator project before it was elevated to top-level project status the following year.&lt;/p&gt;
 &lt;p&gt;Cassandra is a fault-tolerant distributed database that can be used to store and manage large amounts of data across a cluster consisting of numerous commodity servers. The software &lt;a href="https://www.techtarget.com/searchdisasterrecovery/definition/data-replication"&gt;replicates data&lt;/a&gt; across multiple server nodes to avoid single points of failure, and it can be scaled dynamically by adding more servers to a cluster as processing demand increases. Cassandra is designed around distributed scalability and &lt;a href="https://www.techtarget.com/searchdatacenter/definition/high-availability"&gt;high availability&lt;/a&gt;, traditionally with eventual consistency tradeoffs, though Apache Cassandra 5.0 has added transaction capabilities that expand its support for transactional workloads.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Common use cases:&lt;/b&gt; Cassandra is designed for uses that require fast performance, scalability and high availability. It's deployed for various applications, including inventory management, e-commerce, social media analytics, messaging systems and telecommunications, among others.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Licensing:&lt;/b&gt; The Cassandra software is covered by the Apache License 2.0.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Source code repository&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; &lt;a target="_blank" href="https://github.com/apache/cassandra/tree/trunk" rel="noopener"&gt;https://github.com/apache/cassandra/tree/trunk&lt;/a&gt;&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Commercial support options&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; Multiple vendors provide commercial support for Cassandra and DBaaS versions of the database, including DataStax, Aiven and Instaclustr. Amazon Keyspaces (for Apache Cassandra) and Azure Managed Instance for Apache Cassandra are also available as database services from AWS and Microsoft, respectively.&lt;/p&gt;
&lt;/section&gt;       
&lt;section class="section main-article-chapter" data-menu-title="7. Apache CouchDB"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;7. Apache CouchDB&lt;/h2&gt;
 &lt;p&gt;CouchDB is a NoSQL document database first released in 2005 by software engineer Damien Katz that became an Apache project in 2008. The &lt;i&gt;Couch&lt;/i&gt; part of the name is an acronym for "cluster of unreliable commodity hardware," which stems from the project's original goal: to create a reliable database system that could run efficiently on ordinary hardware. CouchDB can be deployed on a single server node, but also as a single logical system across multiple nodes in a cluster, which can be scaled as needed by adding more servers.&lt;/p&gt;
 &lt;p&gt;The database uses JSON documents to store data and JavaScript as its query language. Other key features include support for MVCC and the ACID properties in individual documents, although an eventual consistency model is used for data stored on multiple database servers -- a tradeoff that prioritizes availability and performance over absolute data consistency. Data is synchronized across servers through an incremental replication feature that can be set up for bidirectional tasks and used to support mobile apps and other offline-first applications.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Common use cases:&lt;/b&gt; CouchDB is used for various purposes, including data analytics, time series data storage and mobile applications that require offline storage and functionality.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Licensing: &lt;/b&gt;CouchDB is licensed under the Apache License 2.0.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Source code repository&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; &lt;a target="_blank" href="https://github.com/apache/couchdb" rel="noopener"&gt;https://github.com/apache/couchdb&lt;/a&gt;&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Commercial support options: &lt;/b&gt;The IBM Cloudant cloud database is based on CouchDB with added open source technology that supports full-text search and geospatial indexing. Several other companies also offer support for CouchDB, including packaged instances in the AWS, Azure and Google clouds.&lt;/p&gt;
&lt;/section&gt;       
&lt;section class="section main-article-chapter" data-menu-title="8. Neo4j"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;8. Neo4j&lt;/h2&gt;
 &lt;p&gt;Neo4j is a NoSQL graph database that's well-suited for representing and querying highly connected data sets. Neo4j uses a property graph database model consisting of nodes, which represent individual data entities, and relationships -- also referred to as edges -- that define how different nodes are organized and connected. Nodes and relationships can also include properties, or attributes, in the form of key-value pairs that further describe them.&lt;/p&gt;
 &lt;p&gt;First released as open source software in 2007, Neo4j is overseen by database vendor Neo4j Inc. It originally was a solely Java-based graph database, but has since been expanded to include additional capabilities, including &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Vector-search-now-a-critical-component-of-GenAI-development"&gt;vector search and data storage&lt;/a&gt;. Key features include full ACID compliance, horizontal scalability through an Autonomous Clustering architecture and the Cypher query language. Neo4j Inc. is aligning Cypher with GQL, the ISO graph query language published in April 2024 that uses syntax based on both SQL and Cypher, while continuing to support Cypher extensions not yet included in the standard.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Common use cases:&lt;/b&gt; Typical uses for Neo4j include social networking, recommendation engines, network and IT operations, fraud detection, and supply chain management, with &lt;a href="https://www.techtarget.com/searchenterpriseai/definition/generative-AI"&gt;generative AI&lt;/a&gt; applications also now supported through the vector search feature.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Licensing&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; Neo4j Community Edition is licensed under the GPL version 3. An open source version of Cypher named openCypher is also available under the Apache License 2.0.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Source code repository&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; &lt;a target="_blank" href="https://github.com/neo4j/neo4j" rel="noopener"&gt;https://github.com/neo4j/neo4j&lt;/a&gt;&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Commercial support options:&lt;/b&gt; Neo4j Inc. provides several supported commercial offerings, including&amp;nbsp; Neo4j Enterprise Edition with added closed source components and the subscription-based Neo4j AuraDB cloud service.&lt;/p&gt;
&lt;/section&gt;       
&lt;section class="section main-article-chapter" data-menu-title="9. Couchbase Server"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;9. Couchbase Server&lt;/h2&gt;
 &lt;p&gt;Couchbase Server is a NoSQL document database with multimodel capabilities for storing data both in JSON documents and as key-value pairs. The technology resulted from the 2011 merger of two open source database companies: CouchOne, which had been founded by CouchDB creator Damien Katz to offer systems based on that database, and Membase, which was set up to build a key-value store by developers of the memcached distributed caching technology. The combined company became Couchbase, leading to the development of Couchbase Server.&lt;/p&gt;
 &lt;p&gt;Despite their similar names and partly shared origins, Couchbase Server and CouchDB aren't directly related or compatible -- they're different database technologies with their own code and APIs. Couchbase Server supports strong consistency, distributed ACID transactions and SQL++, a SQL-like language for querying JSON data. It also includes vector and full-text search, plus a multidimensional scaling feature that enables different database functions to be isolated and independently scaled based on workload demands.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Common use cases&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; Couchbase Server is often used to support distributed application workloads and for mobile, edge and IoT applications.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Licensing:&lt;/b&gt; Originally available under the Apache License 2.0, Couchbase Server was switched in 2021 to the Business Source License (BSL) 1.1, a source-available license that restricts commercial use of the software by other vendors. Database releases are converted back to the Apache open source license four years after they become available.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Source code repository&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; &lt;a target="_blank" href="https://github.com/couchbase/manifest" rel="noopener"&gt;https://github.com/couchbase/manifest&lt;/a&gt;&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Commercial support options:&lt;/b&gt; Couchbase offers an enterprise edition of Couchbase Server for cloud and on-premises deployments, as well as a mobile version of the database and a fully managed DBaaS technology named Couchbase Capella.&lt;/p&gt;
&lt;/section&gt;       
&lt;section class="section main-article-chapter" data-menu-title="10. MongoDB"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;10. MongoDB&lt;/h2&gt;
 &lt;p&gt;&lt;a href="https://www.techtarget.com/searchdatamanagement/definition/MongoDB"&gt;MongoDB&lt;/a&gt; is another NoSQL document database that was initially developed as open source software and is now a source-available technology. First released in 2009, MongoDB stores data in a JSON-like document format called BSON, which is short for Binary JSON. As the name indicates, BSON encodes data in a binary structure that's designed to support more data types and faster indexing and querying performance than JSON.&lt;/p&gt;
 &lt;p&gt;The database is often seen as an attractive option for developers who want to build applications without the constraints of a fixed schema. In addition to its document data model, MongoDB includes native support for graph, geospatial and time series data. MongoDB Atlas, a cloud database service offered by lead developer MongoDB Inc., also provides vector and full-text search features that can be used free of charge for development and testing in local environments. Other key features in MongoDB include multi-document ACID transactions, sharding for horizontal scalability and automatic load balancing.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Common use cases:&lt;/b&gt; MongoDB is widely deployed for uses that include AI, edge computing, IoT, mobile, payment and gaming applications, as well as website personalization, content management and product catalogs.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Licensing:&lt;/b&gt; Since 2018, new versions of MongoDB Community Server and patches for previous releases have been made available under the Server Side Public License (SSPL) Version 1, a source available license created by MongoDB Inc.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Source code repository&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; &lt;a target="_blank" href="https://github.com/mongodb/mongo" rel="noopener"&gt;https://github.com/mongodb/mongo&lt;/a&gt;&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Commercial support options: &lt;/b&gt;In addition to MongoDB Atlas, MongoDB Inc. offers a self-managed MongoDB Enterprise Server that also provides additional capabilities beyond what's in the community edition. MongoDB support and managed services are also available from vendors such as Datavail and Percona. Amazon DocumentDB (with MongoDB compatibility) is a fully managed DBaaS offering from AWS that supports MongoDB versions 6.0, 7.0 and 8.0 as of November 2025.&lt;/p&gt;
&lt;/section&gt;       
&lt;section class="section main-article-chapter" data-menu-title="11. Redis"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;11. Redis&lt;/h2&gt;
 &lt;p&gt;Redis is a NoSQL in-memory database that was converted to a source-available technology in March 2024. The Redis project was created in 2009 by software programmer Salvatore Sanfilippo, known by the nickname “antirez,” to help solve a database scaling problem with a real-time website log analysis tool. Short for Remote Dictionary Server, Redis originally was positioned as software that provided a key-value data store as a caching technology to accelerate existing databases and application workloads.&lt;/p&gt;
 &lt;p&gt;The database caching functionality remains the foundation of Redis, with features that include built-in replication, on-disk data persistence and support for complex data types. But the platform has been expanded to include additional capabilities, such as support for &lt;a href="https://www.techtarget.com/searchdatamanagement/news/252514648/Redis-launches-JSON-database-capabilities-with-RedisJSON-20"&gt;storing JSON documents &lt;/a&gt;and both vector and time series data. A graph database module was also added, but lead developer Redis Inc. stopped developing it in 2023.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Common use cases:&lt;/b&gt; While Redis can be used as a full database, one of its most common uses is still as a database query caching layer. It's also often used to support real-time notifications through an integrated pub/sub capability and as a session store to help manage user sessions for web and mobile applications.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Licensing&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; As of March 2026, Redis licensing now varies by version. Redis 8 is available under a tri-license of the Redis Source Available License 2.0, SSPL v1 and AGPLv3. Earlier Redis Community Edition releases from 7.4 through 7.8 were dual-licensed under RSALv2 and SSPLv1, while Redis 7.2 and earlier were released under the BSD 3-Clause license.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Source code repository&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; &lt;a target="_blank" href="https://github.com/redis/redis" rel="noopener"&gt;https://github.com/redis/redis&lt;/a&gt;&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Commercial support options:&lt;/b&gt; Redis offers Redis Software for self-managed deployments and Redis Cloud as its fully managed service. Microsoft offers Azure Cache for Redis as a managed service, but has announced a retirement timeline for all SKUs by 2028 and recommends migration to Azure Managed Redis. Redis managed services are also available from Aiven and Instaclustr, while AWS, Google and Oracle offer cloud services with Redis compatibility.&lt;/p&gt;
&lt;/section&gt;       
&lt;section class="section main-article-chapter" data-menu-title="12. CockroachDB"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;12. CockroachDB&lt;/h2&gt;
 &lt;p&gt;CockroachDB is a source-available distributed SQL database loosely inspired by Google's proprietary Spanner database. Developed primarily by vendor Cockroach Labs, CockroachDB was first released in 2015, with an initial production version appearing two years later. Just like the insect it's named after, a core design goal of CockroachDB is to be hard to kill. The cloud-native database is built to be a fault-tolerant, resilient and consistent &lt;a href="https://www.techtarget.com/searchdatamanagement/definition/data-management"&gt;data management&lt;/a&gt; platform.&lt;/p&gt;
 &lt;p&gt;CockroachDB scales horizontally and can survive various types of equipment failures with minimal disruptions to users and no manual intervention required by DBAs, according to its developers. Key features include automated repair and recovery, support for ACID transactions with strong consistency, a SQL API and geo-partitioning of data to boost application performance. It also has a "multi-active availability" model that enables users to read and write data from any cluster node with no conflicts.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Common use cases: &lt;/b&gt;CockroachDB is well suited for high-volume OLTP applications and distributed database deployments across multiple data centers and geographic regions.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Licensing&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; Since 2019, most of CockroachDB's core features have been licensed under a version of the BSL that requires other vendors to purchase a license from Cockroach Labs if they want to offer a commercial database service. Other core features are covered by the Cockroach Community License (CCL), which allows source code to be viewed and modified but not reused without an agreement on that with Cockroach Labs. The features licensed under the BSL convert to the Apache License 2.0 and become open source 3 years after a new database release, a change that doesn't apply to the CCL-licensed features.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Source code repository&lt;/b&gt;&lt;b&gt;:&lt;/b&gt; &lt;a target="_blank" href="https://github.com/cockroachdb/cockroach" rel="noopener"&gt;https://github.com/cockroachdb/cockroach&lt;/a&gt;&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Commercial support options:&lt;/b&gt; Cockroach Labs provides technical support and additional paid enterprise features that are available in both self-managed and DBaaS deployments.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Editor's note:&lt;/b&gt; This article was originally published in May 2024 and republished in March 2026 to include product updates and improve the reader experience.&lt;/p&gt;
 &lt;p&gt;&lt;i&gt;Sean Michael Kerner is an IT consultant, technology enthusiast and tinkerer. He has pulled Token Ring, configured NetWare and been known to compile his own Linux kernel. He consults with industry and media organizations on technology issues.&lt;/i&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>Open source databases offer viable alternatives to proprietary systems. This guide offers 12 popular options across relational, NoSQL and source-available technologies.</description>
            <image>https://cdn.ttgtmedia.com/rms/onlineimages/code_g1255337870.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/feature/Top-open-source-databases-to-consider</link>
            <pubDate>Wed, 18 Mar 2026 09:00:00 GMT</pubDate>
            <title>12 top open source databases for enterprise use</title>
        </item>
        <item>
            <body>&lt;p&gt;A one‑size‑fits‑all approach to AI models in data-driven applications risks leaving enterprises paying more, waiting longer for analytics results and taking on avoidable business risk. &amp;nbsp;&lt;/p&gt; 
&lt;p&gt;For data and analytics leaders, the era of "bigger is better" in AI applications is giving way to right-sizing strategies that assign small language models (SLMs) to narrow, repeatable tasks and reserve large language models (LLMs) for complex reasoning workloads. The &lt;a href="https://www.techtarget.com/searchenterpriseai/tip/Why-small-language-models-are-on-the-rise"&gt;advantages of SLMs&lt;/a&gt; include easier governance, lower run-rate costs and the ability to run on-premises, which is particularly attractive to regulated industries where compliance and data privacy are paramount. Analyst outlooks and vendor guidance show organizations increasingly moving toward this mixed-model configuration.&lt;/p&gt; 
&lt;p&gt;"There is a space for smaller models to be utilized more," said FTI Consulting managing director and chief data scientist Dimitris Korres. He noted that organizations "can use a mix of large and small models to reduce costs and get better, faster outputs for certain tasks."&lt;/p&gt; 
&lt;div class="youtube-iframe-container"&gt;
 &lt;iframe id="ytplayer-0" src="https://www.youtube.com/embed/AlwWuSor_M4?autoplay=0&amp;amp;modestbranding=1&amp;amp;rel=0&amp;amp;widget_referrer=null&amp;amp;enablejsapi=1&amp;amp;origin=https://www.techtarget.com" type="text/html" height="360" width="640" frameborder="0"&gt;&lt;/iframe&gt;
&lt;/div&gt; 
&lt;section class="section main-article-chapter" data-menu-title="Small models are on the rise"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Small models are on the rise&lt;/h2&gt;
 &lt;p&gt;This &lt;a href="https://www.techtarget.com/searchenterpriseai/opinion/Small-language-models-are-taking-the-spotlight"&gt;move to rightsize models&lt;/a&gt; to make them fit for purpose rather than simply using the largest model emerged as a viable strategy in 2024. It got increased attention when Nvidia researchers published a September 2025 paper titled &lt;i&gt;Small Language Models are the Future of Agentic AI.&lt;/i&gt;&lt;/p&gt;
 &lt;p&gt;In it, they argued that SLMs are "sufficiently powerful, inherently more suitable, and necessarily more economical" for many agentic AI workloads and recommended the use of heterogeneous agents that send low-level tasks to an SLM and only escalate more advanced work to an LLM.&lt;/p&gt;
 &lt;p&gt;Interest in SLMs is still relatively limited, as most organizations remain focused on how to use the dominant LLMs to transform their workflows, products and services.&lt;/p&gt;
 &lt;p&gt;Yet, there is evidence of increasing use of SLMs, signifying a strategic shift toward task‑specific models. Gartner projects that by 2027, organizations will use SLMs three times more than general‑purpose LLMs. A 2025 SNS Insider report valued the global SLM market at $7.9 billion in 2023, predicting it will reach $29.64 billion by 2032.&lt;/p&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="Map tasks to the model size"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Map tasks to the model size&lt;/h2&gt;
 &lt;p&gt;LLMs have tens of billions to hundreds of billions of parameters. This scale makes LLMs valuable, enabling them to work through &lt;a href="https://www.techtarget.com/searchenterpriseai/tip/How-to-evaluate-LLMs-for-enterprise-use-cases"&gt;large numbers of complex, broad tasks&lt;/a&gt;. But LLMs require massive amounts of compute power, making them resource-intensive. Due to their scale, many enterprises use LLMs as an endpoint from a cloud provider rather than deploying them internally.&lt;/p&gt;
 &lt;p&gt;SLMs typically max out at 10 billion parameters and usually have far less -- in the millions in many cases. The smaller size of SLMs often makes them faster to serve up results and more feasible to deploy on-premises or in &lt;a href="https://www.techtarget.com/searchcio/tip/Top-edge-computing-trends-to-watch-in-2020"&gt;edge computing environments&lt;/a&gt; for specialized workloads.&lt;/p&gt;
 &lt;p&gt;SLMs have several advantages compared to LLMs:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;&lt;b&gt;Fit for purpose.&lt;/b&gt; Due to their size and speed, SLMs are ideal for simpler work, such as document sorting and summarizing meeting notes.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Better responsiveness.&lt;/b&gt; Smaller models can return answers faster on focused tasks because they require fewer computations.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Lower costs.&lt;/b&gt; Nvidia claims AI inference can be 10 to 30 times less expensive with SLMs than LLMs if the hardware is fully utilized. Using SLMs can also help avoid the higher energy consumption and costs associated with LLMs.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Private by design.&lt;/b&gt; SLMs are especially suited for working with sensitive data -- for example, processing legal contracts, medical records or financial data -- because they can run on in-house servers or edge devices. This keeps the data within the organization's environment rather than in a public cloud.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Consistent results.&lt;/b&gt; SLMs are generally easier to adjust for predictable outputs.&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;p&gt;For data leaders, the takeaway is simple: Start small and then move up to an LLM when needed. This approach helps companies maintain service quality while keeping runtime costs predictable. For high‑volume automation, using SLMs provides steadier spending than running large, cloud‑dependent models.&lt;/p&gt;
 &lt;p&gt;Such factors have prompted Zach Rossmiller, CIO at the University of Montana, to consider where smaller models might be a better fit than LLMs.&lt;/p&gt;
 &lt;p&gt;"It comes down to use cases for us," he said.&lt;/p&gt;
 &lt;p&gt;He has looked at using SLMs for workloads that &lt;a href="https://www.techtarget.com/searchenterpriseai/tip/Address-top-AI-privacy-concerns-with-this-checklist"&gt;keep sensitive data protected&lt;/a&gt; in on-premises systems and to support a possible digital tutoring service running on a Raspberry Pi or other end-user devices for students in rural areas that lack reliable high-speed internet access.&lt;/p&gt;
 &lt;p&gt;Rossmiller said SLMs seem like a good choice for specific use cases like those, where factors such as data privacy and connectivity concerns may prohibit the use of LLMs.&lt;/p&gt;
&lt;/section&gt;          
&lt;section class="section main-article-chapter" data-menu-title="Too many small models might lead to slowdowns"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Too many small models might lead to slowdowns&lt;/h2&gt;
 &lt;p&gt;Holger Mueller, principal analyst and vice president at Constellation Research, is more skeptical about SLMs. He doesn't see much value in them, saying that users will quickly reach their limits and ask, "That's it? That's all it can do?"&lt;/p&gt;
 &lt;p&gt;Mueller said relying on too many SLMs will also push users into an "integrator" role in which they must find ways to coordinate and integrate multiple small models, which could erase the savings in cost and time. He added that another potential disadvantage is the need for specific &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Experts-share-practices-to-overcome-AI-data-readiness"&gt;training on company data for the SLMs to be effective&lt;/a&gt;.&lt;/p&gt;
 &lt;p&gt;SLMs only seem to be advantageous in a handful of rare circumstances where the scenarios are narrow in scope and remain that way, Mueller said.&lt;/p&gt;
 &lt;p&gt;However, he said if more LLM developers start adding small models to their offerings, "that would be something to pay attention to."&lt;/p&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="Consider the architectural realities for a mixed-model platform"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Consider the architectural realities for a mixed-model platform&lt;/h2&gt;
 &lt;p&gt;While SLMs can run on-premises, their usefulness depends on several factors.&lt;/p&gt;
 &lt;p&gt;Most organizations will continue to rely on LLMs, which require data platforms to support secure, high-speed transfers to cloud AI platforms. However, adding task-focused SLMs might require additional investments to build a distributed, hybrid architecture that keeps sensitive data where it already resides to &lt;a href="https://www.techtarget.com/searchenterpriseai/tip/How-to-navigate-data-sovereignty-for-AI-compliance"&gt;meet compliance and privacy requirements&lt;/a&gt; and lower data transfer costs.&lt;/p&gt;
 &lt;p&gt;Nvidia's researchers said enterprises might need to tailor their ecosystems toward AI agents that can automatically route tasks to the right model.&lt;/p&gt;
 &lt;p&gt;"Cloud is always an option, but there are also self-hosting endpoints that have ridiculously low costs," said Korres. "And if a project requires more control, or isolation of the data flow or data residency, then smaller models can definitely be hosted on local devices."&lt;/p&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="On-premises models offer privacy and predictability"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;On-premises models offer privacy and predictability&lt;/h2&gt;
 &lt;p&gt;Because they are smaller and more focused, SLMs can be easier to manage from a governance, risk and compliance perspective, especially if they run on-premises or in a private cloud. The narrow scope of SLMs helps with transparency, monitoring and auditability, reducing the risk of data exposure compared to large, public cloud models.&lt;/p&gt;
 &lt;p&gt;Furthermore, SLMs running locally can improve reproducibility, meaning "that anyone else can follow the same steps and get the same result," said Jonathan Chang, assistant professor of computer science at Harvey Mudd College. In contrast, reproducibility is not guaranteed with models hosted in public clouds.&lt;/p&gt;
 &lt;p&gt;Korres also noted that because SLMs are often used for narrowly defined tasks, organizations can more easily &lt;a href="https://www.techtarget.com/searchsoftwarequality/tip/Benchmarking-LLMs-A-guide-to-AI-model-evaluation"&gt;evaluate their performance than when using a general-purpose LLM&lt;/a&gt;.&lt;/p&gt;
 &lt;p&gt;However, some risks are similar regardless of the model size.&lt;/p&gt;
 &lt;p&gt;"Models small and large will hallucinate. They can produce outputs that could be misleading. They could leak training data that could mean privacy concerns," said Chang.&lt;/p&gt;
 &lt;p&gt;&lt;iframe title="SLMs vs. LLMs: At a glance" aria-label="Table" id="datawrapper-chart-XGxUf" src="https://datawrapper.dwcdn.net/XGxUf/2/" scrolling="no" frameborder="0" style="width: 0; min-width: 100% !important; border: none;" height="638" data-external="1"&gt;&lt;/iframe&gt;&lt;/p&gt;
 &lt;p&gt; &lt;script type="text/javascript"&gt;window.addEventListener("message",function(a){if(void 0!==a.data["datawrapper-height"]){var e=document.querySelectorAll("iframe");for(var t in a.data["datawrapper-height"])for(var r,i=0;r=e[i];i++)if(r.contentWindow===a.source){var d=a.data["datawrapper-height"][t]+"px";r.style.height=d}}});&lt;/script&gt; &lt;/p&gt;
&lt;/section&gt;        
&lt;section class="section main-article-chapter" data-menu-title="Decision rules for cost vs. performance"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Decision rules for cost vs. performance&lt;/h2&gt;
 &lt;p&gt;As many enterprise technology executives are finding, the costs associated with LLM use can be high and unpredictable, Chang said.&lt;/p&gt;
 &lt;p&gt;Additionally, he pointed out the environmental impact of LLMs.&lt;/p&gt;
 &lt;p&gt;"Bigger models require more resources to run. They take more energy, so they emit more carbon," Chang said.&lt;/p&gt;
 &lt;p&gt;By contrast, SLMs provide lower, more predictable operational costs with less impact on the environment, he said. However, gaining those advantages requires having the &lt;a href="https://www.techtarget.com/searchstorage/opinion/IT-leaders-face-data-infrastructure-gaps-as-AI-workloads-grow"&gt;right infrastructure to run the models.&lt;/a&gt; The financial benefits of running SLMs locally might dissipate if the organization needs to set up a new data center or make other IT investments.&lt;/p&gt;
 &lt;p&gt;Enterprises need to fully consider the strategic use of SLMs to make sure the model choice fits the task.&lt;/p&gt;
 &lt;p&gt;"If it is a well-defined task and it is narrow enough, I see more benefits than risks to using a smaller model," said Korres. &lt;b&gt;"&lt;/b&gt;But if the function is quite general, or you need something that can manage agents or define things on the fly, that's where larger models make a difference."&lt;/p&gt;
 &lt;p&gt;Other experts had similar assessments, noting that the workload should &lt;a target="_blank" href="https://mitsloan.mit.edu/ideas-made-to-matter/looking-ahead-ai-and-work-2026" rel="noopener"&gt;determine&lt;/a&gt; whether to go with a small or large model.&lt;/p&gt;
 &lt;p&gt;"If you need a gigantic amount of data, then an LLM is the right answer," said Chang. "But if you're finding that most of your friction is with individual employees needing to automate small tasks, or you're dealing with sensitive data and want full control from end to end, or you care about reproducibility, then small models are a feasible option."&lt;/p&gt;
 &lt;p&gt;&lt;em&gt;Mary K. Pratt is an award-winning freelance journalist with a focus on covering enterprise IT and cybersecurity management.&lt;/em&gt;&lt;/p&gt;
 &lt;p&gt;&amp;nbsp;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>It doesn't need to be a binary choice. Enterprises are warming up to smaller AI models to meet compliance and cost needs while reserving large models for complex jobs.</description>
            <image>https://cdn.ttgtmedia.com/rms/onlineimages/strategy_g1227074505.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/feature/SLM-vs-LLM-Rightsize-data-architecture-to-optimize-AI-use</link>
            <pubDate>Tue, 17 Mar 2026 09:38:00 GMT</pubDate>
            <title>SLM vs. LLM: Rightsize data architecture to optimize AI use</title>
        </item>
        <item>
            <body>&lt;p&gt;AI is in production across the enterprise, but have you adapted your data governance practices to keep pace?&lt;/p&gt; 
&lt;p&gt;As business units chase their own priorities, data remains scattered and inconsistent, heightening the risk of faulty AI output and poor decisions. The fix is not another dashboard but &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/AI-data-governance-guidance-that-gets-you-to-the-finish-line"&gt;a federated data and AI governance model&lt;/a&gt; that establishes ownership, checks accuracy and monitors data quality and lineage in real time. By implementing a combination of technological and organizational changes tailored to use AI at scale, companies can move faster without sacrificing security or compliance.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="Why production AI raises the bar on data governance"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Why production AI raises the bar on data governance&lt;/h2&gt;
 &lt;p&gt;Today, AI has become something organizations depend on, with &lt;a href="https://www.techtarget.com/searchcio/feature/Pillars-of-an-agentic-AI-strategy"&gt;AI agents embedded in workflows&lt;/a&gt; to support task coordination and faster decision-making. The larger the organization, the larger the challenge, as operating companies, business units and departments have their own priorities and data practices. This makes proactive AI data governance a must.&lt;/p&gt;
 &lt;p&gt;For example, say you want to answer a seemingly simple question like, "Who are my most profitable customers?" To do so, you need to be able to gather data from across the enterprise on customers, their purchases and the cost of marketing and selling to them. This helps confirm operations within complex multinational units. Rather than hunt for insights in charts, business users can use conversational analytics and agentic BI to fetch the information they need or explain the context of data.&lt;/p&gt;
 &lt;p&gt;The landscape has shifted in recent years because organizations can no longer completely trust data or assume a person generated it. Enterprises face new risks if their AI tools produce faulty answers after training on outdated or inaccurate models. To combat this issue, many are adopting a zero-trust approach to AI data governance to establish authentication and verification measures.&lt;/p&gt;
 &lt;p&gt;It is estimated that unstructured data typically &lt;a href="https://www.techtarget.com/searchbusinessanalytics/feature/Ng-Biggest-benefit-of-AI-may-be-unlocking-unstructured-data"&gt;takes up about 80% of the enterprise's information assets&lt;/a&gt; and grows about four times faster than structured data. It requires significant effort to align data collection, creation, classification, formatting and use across organizational boundaries. As a result, modern governance tools have evolved into the control plane for handling data at this scale.&lt;/p&gt;
 &lt;div class="imagecaption alignLeft"&gt;
  &lt;img src="https://www.techtarget.com/rms/onlineImages/data_management-need_to_govern_data-f.png" alt="An educational graphic titled "&gt;The key benefits of data governance in organizations include improved data quality, regulatory compliance and more accurate decision-making.
 &lt;/div&gt;
&lt;/section&gt;      
&lt;section class="section main-article-chapter" data-menu-title="Build a federated governance strategy that scales"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Build a federated governance strategy that scales&lt;/h2&gt;
 &lt;p&gt;Data governance is no longer just an IT problem. Today, it's being led by the C-suite in a federated model that expands data governance to include AI governance, covering model inventory, risk assessment and post-deployment monitoring, rather than treating AI as a separate track.&lt;/p&gt;
 &lt;p&gt;Gartner predicts that by 2028, about half of organizations will adopt zero-trust data governance and recommends appointing a dedicated AI governance leader to oversee policies, AI risk management and compliance with data and analytics teams.&lt;/p&gt;
 &lt;p&gt;In the federated model, a steering committee of senior executives and data owners sets policy and resolves definition disputes. This board can require business units to align their systems and processes to approved standards. Distributed data stewards embed the policies into daily workflows.&lt;/p&gt;
 &lt;p&gt;Modern data governance also depends on the platform foundation. As organizations break down operational and analytical silos, the &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Building-a-strong-data-analytics-platform-architecture"&gt;data lakehouse has emerged as the single governed foundation&lt;/a&gt; for analytics and AI, reducing copies and data fragmentation. When paired with active metadata management, the data governance stack alerts teams when data requires recertification or updating.&lt;/p&gt;
 &lt;p&gt;Once that structure is in place, the question becomes how to realize the benefits for a collaborative AI-based organization.&lt;/p&gt;
 &lt;p&gt;These programs also have a clear mission statement, a business case tied to AI operational readiness, training that improves data and AI fluency, and a process for communicating progress and results. The steps outlined here will put organizations on the path to better data governance and more reliable AI.&lt;/p&gt;
 &lt;div class="youtube-iframe-container"&gt;
  &lt;iframe id="ytplayer-0" src="https://www.youtube.com/embed/BqdPuwvwPk4?autoplay=0&amp;amp;modestbranding=1&amp;amp;rel=0&amp;amp;widget_referrer=null&amp;amp;enablejsapi=1&amp;amp;origin=https://www.techtarget.com" type="text/html" height="360" width="640" frameborder="0"&gt;&lt;/iframe&gt;
 &lt;/div&gt;
&lt;/section&gt;        
&lt;section class="section main-article-chapter" data-menu-title="Why the federated governance model pays off"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Why the federated governance model pays off&lt;/h2&gt;
 &lt;p&gt;Here are the key benefits a &lt;a href="https://www.techtarget.com/searchdatamanagement/definition/data-governance"&gt;successful data governance program&lt;/a&gt; can produce in an organization.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;1. Greater efficiency. &lt;/b&gt;With access to well-governed data, AI agents improve operational efficiency across many areas. They take on manual work, such as schema mapping and duplicate merging, giving teams more time to weed out underperforming product lines and invest more in those that show promise. Analyzing business processes can reveal opportunities to improve them -- but only if the data underlying those processes is reliable. According to Gartner, 60% of organizations without cohesive data governance won't get full value out of AI until 2027.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;2. Better data quality.&lt;/b&gt; Despite significant IT spending, data quality problems persist. A 2026 LeBow College of Business study cites ongoing data‑integrity gaps, saying better quality was a priority for 51% of data and analytics leaders. In the same study, 43% said data readiness was the biggest obstacle to AI initiatives. &lt;a href="https://www.techtarget.com/searchenterpriseai/feature/9-data-quality-issues-that-can-sideline-AI-projects"&gt;Improved data quality reduces AI errors&lt;/a&gt; and protects the organization in the event of a breach. An IBM study from 2025 found that the average cost of a data breach in the U.S. was $10.22 million. Well-cataloged data reduces sensitive information and helps security by making it easier to run discovery, classification, access control and encryption.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;3. Better compliance. &lt;/b&gt;The EU AI Act sets obligations on AI systems while GDPR and California's &lt;a href="https://www.techtarget.com/searchsecurity/tip/State-of-data-privacy-laws"&gt;CCPA/CPRA regulate&lt;/a&gt; the collection, use and auditing of personal data, imposing additional compliance requirements across sectors. About 20 U.S. states have enacted comprehensive privacy laws. European regulators imposed €1.2 billion in GDPR fines in 2025. With sums of this size, organizations need accurate, auditable reporting and governance-backed security and privacy controls to reduce the risk of fines and legal action.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;4. Better decision-making.&lt;/b&gt; Sound data gives executives and their teams confidence to make better business decisions on price adjustments, product strategy, customer service and other aspects of operations. This &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Metadata-management-standards-examples-that-guide-success"&gt;depends heavily on metadata management&lt;/a&gt; -- the catalogs that handle governance and notify users when data requires correction -- to ensure accurate data for strategic planning, business intelligence and advanced analytics.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;5. Improved business performance.&lt;/b&gt; Ultimately, the benefits described above should lead to higher revenue and profits as companies rewire their operating models to take advantage of AI capabilities. LeBow's 2026 study &lt;a target="_blank" href="https://www.lebow.drexel.edu/sites/default/files/2026-01/lebow-precisely-state-data-integrity-ai-readiness-2026.pdf" rel="noopener"&gt;found&lt;/a&gt; higher data trust in organizations with governance programs than those without at 71% vs. 50%. Business leaders now treat data and AI literacy as a basic requirement and push for integrating data and AI governance to quicken decision-making and improve business performance.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;6. Enhanced business reputation.&lt;/b&gt; In addition to tangible financial gains, effective data governance produces high-quality data, which fosters better customer interactions and drives higher satisfaction and loyalty.&lt;/p&gt;
 &lt;p&gt;&lt;b&gt;Editor's note:&lt;/b&gt; &lt;i&gt;TechTarget editors updated this article in March 2026 for timeliness and to add new information.&lt;/i&gt;&lt;/p&gt;
 &lt;p&gt;&lt;i&gt;Tom Walat is an editor and reporter for TechTarget, where he covers data technologies.&lt;/i&gt;&lt;/p&gt;
 &lt;p&gt;&lt;i&gt;Andy Hayler is an independent analyst on enterprise data management strategy.&lt;/i&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>Follow this practical blueprint to adopt a modern data governance approach that aligns people, processes and platform to deliver measurable results from AI across the business.</description>
            <image>https://cdn.ttgtmedia.com/visuals/searchContentManagement/governance_strategy/contentmanagement_article_009.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/tip/5-benefits-of-building-a-strong-data-governance-strategy</link>
            <pubDate>Tue, 10 Mar 2026 09:15:00 GMT</pubDate>
            <title>Build trust on a federated governance model</title>
        </item>
        <item>
            <body>&lt;p&gt;When data sprawls across various repositories, &lt;a href="https://www.techtarget.com/searchdatamanagement/definition/data-management"&gt;data management&lt;/a&gt; becomes more challenging. Analytics and AI applications are also less effective if data scientists and other end users can't find relevant data or understand its business context. In many cases, "organizations are drowning in data yet starving for insights," said Priya Iragavarapu, managing director of the AI practice at consulting firm AArete.&lt;/p&gt; 
&lt;p&gt;Data catalogs provide a unified inventory of enterprise data assets, making them more manageable, accessible and understandable. Data management teams can use a wide range of tools to &lt;a href="https://www.techtarget.com/searchdatamanagement/answer/What-steps-are-key-to-building-a-data-catalog"&gt;build and manage catalogs&lt;/a&gt;. Data catalog tools collect metadata from various data sources and use it to organize, classify and enrich data entries. They're commonly integrated with &lt;a href="https://www.techtarget.com/searchdatamanagement/definition/data-governance"&gt;data governance&lt;/a&gt; software to help organizations manage data quality, data use and regulatory compliance.&lt;/p&gt; 
&lt;p&gt;Increasingly, data catalog software also incorporates generative AI (GenAI), machine learning (ML) and other AI technologies to streamline catalog development and use. For example, early data catalogs required custom scripts to crawl data and harvest metadata, but modern tools do so automatically. AI assistants and agents handle cataloging tasks and help end users find data.&lt;/p&gt; 
&lt;p&gt;To help inform product evaluations by data leaders, the following are 15 notable data catalog tools, listed in alphabetical order with details on their key features and capabilities. TechTarget editors compiled the list based on research of available technologies, as well as market reports and vendor rankings from Forrester Research and Gartner.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="1. Alation Data Catalog"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;1. Alation Data Catalog&lt;/h2&gt;
 &lt;p&gt;Alation Data Catalog uses AI, ML, automation and natural language processing to simplify data discovery, create business glossaries and power its core Behavioral Analysis Engine. The engine generates popularity rankings, usage recommendations and other insights about data sets. It also analyzes data usage patterns to help streamline &lt;a href="https://www.techtarget.com/searchdatamanagement/definition/data-stewardship"&gt;data stewardship&lt;/a&gt;, data governance and query optimization processes.&lt;/p&gt;
 &lt;p&gt;Allie AI is an AI copilot that documents new data assets, recommends metadata descriptions and identifies potential data stewards. Alation Data Catalog includes a set of prebuilt analytics dashboards with customizable reporting, and collaboration features enable users to create wiki articles and searchable conversations in data catalogs. The tool is part of the broader Alation Agentic Data Intelligence Platform, which also offers data governance, &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/How-data-lineage-tools-boost-data-governance-policies"&gt;data lineage&lt;/a&gt; and data product marketplace applications.&lt;/p&gt;
 &lt;p&gt;Other key features in Alation Data Catalog include the following:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Capabilities for flagging data health issues and defining enterprise &lt;a href="https://www.techtarget.com/searchcio/definition/data-governance-policy"&gt;data governance policies&lt;/a&gt;.&lt;/li&gt; 
  &lt;li&gt;More than 120 connectors to data sources, plus an Open Connector Framework SDK for building custom ones.&lt;/li&gt; 
  &lt;li&gt;A SQL editor for creating data queries that can be published in catalogs for sharing and reuse.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="2. Alex Augmented Data Catalog"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;2. Alex Augmented Data Catalog&lt;/h2&gt;
 &lt;p&gt;Alex Augmented Data Catalog provides various automation, AI and ML capabilities to support catalog creation and metadata management. Developed by Alex Solutions, the software automates data discovery and cataloging. It includes built-in features for data profiling, lineage tracking and metadata enrichment, plus a set of AI agents for tasks such as &lt;a href="https://www.techtarget.com/searchsecurity/tip/How-to-write-a-data-classification-policy-with-template"&gt;data classification&lt;/a&gt; and anomaly detection in data sets.&lt;/p&gt;
 &lt;p&gt;The data catalog tool also automates aspects of data governance and &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Proactive-practices-for-data-quality-improvement"&gt;data quality processes&lt;/a&gt;. Data governance managers can use it to create policies, assign data stewards and monitor compliance with internal policies and regulatory requirements. The software automatically identifies data quality issues, and data stewards can analyze their potential impact on business workflows and alert data owners about necessary fixes.&lt;/p&gt;
 &lt;p&gt;Alex Augmented Data Catalog also provides the following features:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Google-like natural language search and query capabilities.&lt;/li&gt; 
  &lt;li&gt;Plug-and-play metadata connectors to various data sources.&lt;/li&gt; 
  &lt;li&gt;A no-code ontology for classifying and organizing data based on business terminology, processes and objectives.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="3. Ataccama Data Catalog"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;3. Ataccama Data Catalog&lt;/h2&gt;
 &lt;p&gt;Ataccama Data Catalog is a core component of Ataccama One, an AI-driven platform centered on data quality management. The tool automatically monitors data sets for anomalies, data quality issues and structural changes while providing built-in quality rules as well as capabilities for creating custom ones. It also captures data lineage documentation and includes data profiling, data classification and metadata management capabilities.&lt;/p&gt;
 &lt;p&gt;An AI agent added to Ataccama One in November 2025 handles various tasks autonomously in data catalogs, such as profiling data, &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Evaluating-data-quality-requires-clear-and-measurable-KPIs"&gt;assessing data quality&lt;/a&gt;, and creating and applying quality rules. GenAI capabilities enable catalog users to create SQL queries, generate data descriptions and perform other tasks in natural language. In addition, Ataccama Data Catalog runs AI and ML algorithms to identify patterns, trends and relationships in data sets.&lt;/p&gt;
 &lt;p&gt;The catalog software also includes the following features:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Indexing of reports, dashboards and data stories for BI and analytics uses.&lt;/li&gt; 
  &lt;li&gt;Collaboration features that enable users to add comments and ask questions about data assets.&lt;/li&gt; 
  &lt;li&gt;A data marketplace function to make data products available for reuse.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="4. Atlan Data Catalog"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;4. Atlan Data Catalog&lt;/h2&gt;
 &lt;p&gt;Atlan Data Catalog borrows design principles from Google and tools such as GitHub and Slack. It enables users to search for data assets in natural language using keywords or associated business metrics, while providing a SQL syntax search capability for data engineers. Organizations can also integrate collaborative data workflows into catalogs. For example, users can discuss data in Slack chats and create Jira tickets to report data issues.&lt;/p&gt;
 &lt;p&gt;A Companion Sidebar feature provides at-a-glance information about data lineage, usage history, Slack threads, Jira issues and more to help users decide whether data is relevant and trustworthy. Atlan AI, a copilot tool, generates descriptions of data assets, data lineage summaries, SQL queries and definitions of business terms, metrics and KPIs.&lt;/p&gt;
 &lt;p&gt;Part of a broader data and metadata management platform that aims to create an enterprise context layer in organizations, Atlan Data Catalog also includes the following features:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Open APIs that enable fully customizable metadata ingestion.&lt;/li&gt; 
  &lt;li&gt;More than 80 connectors to data platforms and tools.&lt;/li&gt; 
  &lt;li&gt;Role-based filtering to personalize catalog browsing for different users.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="5. AWS Glue Data Catalog"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;5. AWS Glue Data Catalog&lt;/h2&gt;
 &lt;p&gt;AWS Glue Data Catalog is the persistent metadata store in AWS Glue, a fully managed extract, transform and load (ETL) service. Data management teams can use it to store, annotate and share metadata for use in ETL data integration jobs on the AWS cloud platform. It also provides a consistent metadata layer for querying and analyzing data across various AWS data stores, using integrated analytics services such as Amazon Athena, Amazon EMR, Amazon Redshift Spectrum and Amazon SageMaker AI.&lt;/p&gt;
 &lt;p&gt;As in traditional relational database catalogs, AWS Glue Data Catalog organizes metadata into databases and tables. The software is compatible with the metastore repository in Apache Hive and can be used as an external metastore for Hive data in Amazon EMR clusters. Organizations can also import technical metadata from the catalog tool into business data catalogs in Amazon DataZone, a separate data management service.&lt;/p&gt;
 &lt;p&gt;Other features in AWS Glue Data Catalog include the following:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;A wizard for creating crawlers that automatically scan data sources and extract metadata.&lt;/li&gt; 
  &lt;li&gt;Automated schema management and data lineage documentation.&lt;/li&gt; 
  &lt;li&gt;Integration with AWS Lake Formation for defining and managing data access policies.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="6. Coalesce Catalog"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;6. Coalesce Catalog&lt;/h2&gt;
 &lt;p&gt;Known as &lt;i&gt;CastorDoc&lt;/i&gt; before Coalesce acquired and renamed it in March 2025, this AI-powered tool provides automated data documentation and natural language search capabilities. An AI assistant helps catalog users find relevant data, write SQL queries and understand data governance policies. Coalesce Catalog also automatically maps data lineage information and creates a metadata-driven semantic layer that applies business context to the data in a catalog.&lt;/p&gt;
 &lt;p&gt;Advanced data filtering, popularity signals, freshness indicators and certification badges help users evaluate data assets for use in analytics applications. Coalesce Catalog streamlines data classification and incorporates data governance features, including role-based access control and guided access request workflows. The catalog software is offered in a platform alongside Coalesce's original data transformation tool; the company is working to fully integrate the two tools.&lt;/p&gt;
 &lt;p&gt;Coalesce Catalog also includes the following features:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Integration with more than 30 data platforms and related tools.&lt;/li&gt; 
  &lt;li&gt;Interfaces for searching data catalogs in Slack or Teams.&lt;/li&gt; 
  &lt;li&gt;Audit trails for monitoring data use and regulatory compliance.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="7. Collibra Data Catalog"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;7. Collibra Data Catalog&lt;/h2&gt;
 &lt;p&gt;Collibra offers a namesake &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Data-and-AI-governance-must-team-up-for-AI-to-succeed"&gt;data and AI governance&lt;/a&gt; platform centered on Collibra Data Catalog. The tool provides automated data discovery, classification and curation powered by AI and ML, including the use of GenAI to create descriptions of data assets. It also automates data profiling and data lineage mapping across source systems. An integrated AI copilot helps catalog users find data and associated business definitions.&lt;/p&gt;
 &lt;p&gt;Collibra Data Catalog includes more than 100 prebuilt integrations for ingesting metadata from various data stores, business applications, BI platforms and &lt;a href="https://www.techtarget.com/searchbusinessanalytics/feature/15-data-science-tools-to-consider-using"&gt;data science tools&lt;/a&gt;. It provides configurable workflows for managing data catalogs, as well as guided data stewardship features and controls for enforcing data security and privacy protections. An embedded semantic layer connects technical metadata to business terms and concepts.&lt;/p&gt;
 &lt;p&gt;The Collibra software also offers the following features:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Built-in views of &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Data-governance-metrics-Data-quality-data-literacy-and-more"&gt;data quality metrics&lt;/a&gt; and support for certifying trustworthy data.&lt;/li&gt; 
  &lt;li&gt;Collaboration capabilities, including crowdsourced feedback on data assets through ratings, reviews and comments.&lt;/li&gt; 
  &lt;li&gt;An integrated data marketplace where users can search for relevant data products and other curated data assets.&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;figure class="main-article-image full-col" data-img-fullsize="https://www.techtarget.com/rms/onlineimages/example_of_how_a_data_catalog_works-f.png"&gt;
  &lt;img data-src="https://www.techtarget.com/rms/onlineimages/example_of_how_a_data_catalog_works-f_mobile.png" class="lazy" data-srcset="https://www.techtarget.com/rms/onlineimages/example_of_how_a_data_catalog_works-f_mobile.png 960w,https://www.techtarget.com/rms/onlineimages/example_of_how_a_data_catalog_works-f.png 1280w" alt="Diagram showing an example of how data catalogs work." height="319" width="560"&gt;
  &lt;figcaption&gt;
   &lt;i class="icon pictures" data-icon="z"&gt;&lt;/i&gt;Data catalog tools automate the creation and management of data inventories in organizations.
  &lt;/figcaption&gt;
  &lt;div class="main-article-image-enlarge"&gt;
   &lt;i class="icon" data-icon="w"&gt;&lt;/i&gt;
  &lt;/div&gt;
 &lt;/figure&gt;
&lt;/section&gt;      
&lt;section class="section main-article-chapter" data-menu-title="8. Data.world"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;8. Data.world&lt;/h2&gt;
 &lt;p&gt;Acquired by ServiceNow in July 2025, Data.world is a cloud-native data catalog tool offered as a SaaS platform. It's built on a knowledge graph architecture that provides a semantically organized view of enterprise data assets and their associated metadata across disparate systems. It also automates data quality checks, tracks data lineage and creates visualized maps of data relationships and dependencies.&lt;/p&gt;
 &lt;p&gt;Data.world includes a set of AI bots that help organizations deploy and manage data catalogs and automate data governance tasks. Archie Chat, a conversational AI assistant, provides a chat-like data discovery interface to assist catalog users in data searches, suggest research questions and generate natural language descriptions of data assets and metadata.&lt;/p&gt;
 &lt;p&gt;Other notable features in Data.world include the following:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Collaborative querying capabilities and automated documentation of queries and related comments in a searchable repository.&lt;/li&gt; 
  &lt;li&gt;Customizable data governance workflows and task management processes.&lt;/li&gt; 
  &lt;li&gt;A data product marketplace with an online shopping UX.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="9. Dataplex Universal Catalog"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;9. Dataplex Universal Catalog&lt;/h2&gt;
 &lt;p&gt;Dataplex Universal Catalog ingests technical metadata from Google Cloud and on-premises data sources and enables users to enrich it with business context. Google released the tool in 2024 to replace an older data catalog service. New features include a unified web interface and API, more advanced governance capabilities and wider metadata support.&lt;/p&gt;
 &lt;p&gt;Metadata change feeds enable data teams to track metadata updates in near real time and trigger automated workflows, such as data quality scans, compliance audits and security policy updates, when specified changes occur. Dataplex Universal Catalog also integrates with Google's BigQuery data platform and Vertex AI service to support data and &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Exploding-interest-in-GenAI-makes-AI-governance-a-necessity"&gt;AI governance initiatives&lt;/a&gt;. It automatically captures data lineage documentation and includes built-in data profiling and data quality management capabilities.&lt;/p&gt;
 &lt;p&gt;The catalog software also includes the following features:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Automated metadata harvesting from various Google Cloud data sources, and support for ingesting metadata from other systems.&lt;/li&gt; 
  &lt;li&gt;Keyword and natural language search options.&lt;/li&gt; 
  &lt;li&gt;Search-driven access to data insights generated in BigQuery Studio using Google's Gemini AI assistant.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="10. Erwin Data Catalog"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;10. Erwin Data Catalog&lt;/h2&gt;
 &lt;p&gt;Erwin Data Catalog is now part of Quest Trusted Data Management Platform, a software suite introduced in February 2026 that also includes Quest Software's data modeling, governance, quality and marketplace tools. The software automatically harvests, catalogs, enriches and curates metadata. It also supports drag-and-drop data mapping, reference data management, data lifecycle management, data lineage documentation and data classification.&lt;/p&gt;
 &lt;p&gt;Standard data connectors ingest data from commonly used databases. Optional ones are available for streaming data, cloud applications, BI environments and other data sources. Erwin Data Catalog integrates with a companion data literacy tool to aid in data discovery and governance. Built-in version management and change control functions track changes to data mappings and documentation.&lt;/p&gt;
 &lt;p&gt;The catalog tool also provides the following features:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;A dashboard provides high-level views of data catalog metrics and drill-down analysis capabilities.&lt;/li&gt; 
  &lt;li&gt;An impact analysis function for assessing the potential effects of changes to data attributes or tables.&lt;/li&gt; 
  &lt;li&gt;Automated functions accelerate data movement and transformation, as well as code generation and documentation.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="11. IBM Watsonx.data intelligence"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;11. IBM Watsonx.data intelligence&lt;/h2&gt;
 &lt;p&gt;IBM Watsonx.data intelligence is a data governance and metadata management software suite launched in May 2025 that includes the former IBM Knowledge Catalog tool. It catalogs structured, unstructured and semistructured data, as well as ML models and other analytics assets. It supports AI-driven data discovery and provides &lt;a href="https://www.techtarget.com/searchdatamanagement/opinion/Human-oversight-enables-automated-data-governance"&gt;automated data governance&lt;/a&gt; functions for tasks such as data quality assessments and data privacy policy management.&lt;/p&gt;
 &lt;p&gt;The software also includes metadata enrichment capabilities powered by large language models, plus a set of Knowledge Accelerators -- industry-specific vocabularies of business terms designed to streamline data governance and analytics deployments. It visually maps relationships between data assets and governance artifacts, using a knowledge graph and the FoundationDB open source database originally developed by Apple.&lt;/p&gt;
 &lt;p&gt;The catalog tool offers the following features as well:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Data profiling, cleansing and validation capabilities.&lt;/li&gt; 
  &lt;li&gt;Support for creating data protection rules to control access to sensitive data.&lt;/li&gt; 
  &lt;li&gt;Integration with data lineage and data product marketplace tools that are also part of Watsonx.data intelligence.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="12. Informatica Data Catalog"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;12. Informatica Data Catalog&lt;/h2&gt;
 &lt;p&gt;Informatica Data Catalog is part of the AI-powered Intelligent Data Management Cloud (IDMC) platform developed by Informatica, which Salesforce acquired in November 2025. The catalog tool uses Claire, Informatica's AI engine, to automatically find, ingest, classify and inventory data. Automated data curation features also use AI and ML algorithms to identify relationships between data sets and associate business terms with technical metadata.&lt;/p&gt;
 &lt;p&gt;Data lineage capabilities track data as it moves through systems and data pipelines, supporting impact analysis when data changes. Built-in collaboration capabilities let users add reviews, ratings and annotations to data assets, and subject matter experts can answer questions through a Q&amp;amp;A feature. Informatica Data Catalog also integrates with other IDMC tools, including data governance and data marketplace services.&lt;/p&gt;
 &lt;p&gt;In addition, the catalog software provides the following features:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Automated data profiling and built-in functions for applying data quality rules and monitoring quality levels.&lt;/li&gt; 
  &lt;li&gt;A natural language search function and browsable hierarchical views for finding relevant data in a catalog.&lt;/li&gt; 
  &lt;li&gt;A knowledge graph that visually displays the connections between related data assets.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="13. Microsoft Purview Unified Catalog"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;13. Microsoft Purview Unified Catalog&lt;/h2&gt;
 &lt;p&gt;This tool is part of Microsoft Purview, a data security, governance and compliance service that runs in the Microsoft Azure cloud. Initially known as &lt;i&gt;Microsoft Purview Data Catalog&lt;/i&gt;, it was renamed in late 2024, when Microsoft launched a revised data governance offering. The software runs on top of Microsoft Purview Data Map, a companion metadata management tool that scans data sources, ingests metadata and automatically classifies data.&lt;/p&gt;
 &lt;p&gt;In Microsoft Purview Unified Catalog, users can search for individual data assets or data products, such as tables, files and Microsoft Power BI reports. An embedded business glossary can also be used to find relevant data products by searching for terms, key data elements or business objectives. An AI copilot aids in catalog searches.&lt;/p&gt;
 &lt;p&gt;Other features in the catalog tool include the following:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Data curation for organizing data by governance domains and grouping related data assets and products.&lt;/li&gt; 
  &lt;li&gt;Built-in data quality rules, plus data quality scanning, scoring and alerting functions.&lt;/li&gt; 
  &lt;li&gt;Workflows to help organizations track data governance practices and address issues.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="14. OvalEdge Data Catalog"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;14. OvalEdge Data Catalog&lt;/h2&gt;
 &lt;p&gt;OvalEdge Data Catalog is the foundation of OvalEdge's namesake data governance platform. The catalog software crawls data sources and uses AI and ML algorithms to ingest, organize, enrich and curate data assets. It also provides AI-driven data classification and automated data lineage generation, including data flow diagrams that show how data moves through systems. Users can create custom fields to collect extended metadata types, such as access permissions and source-specific attributes.&lt;/p&gt;
 &lt;p&gt;OvalEdge Data Catalog includes built-in functions for data profiling and documenting relationships between data objects. It also tracks data use and generates popularity and importance scores to help teams prioritize data curation efforts. The tool supports both keyword and natural language search. OvalEdge's agentic AI chatbot, askEdgi, streamlines metadata search and analysis and triggers automated data governance workflows.&lt;/p&gt;
 &lt;p&gt;The data catalog software also includes the following features:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Native connectors to more than 150 data sources.&lt;/li&gt; 
  &lt;li&gt;Question Wall, a centralized hub for knowledge sharing and collaboration.&lt;/li&gt; 
  &lt;li&gt;Integration with tools such as Slack, Jira and ServiceNow.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="15. Precisely Data Catalog"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;15. Precisely Data Catalog&lt;/h2&gt;
 &lt;p&gt;A foundational component of Precisely Data Integrity Suite -- a broad data management and governance platform -- Precisely Data Catalog uses AI algorithms to automatically ingest metadata and generate data descriptions. Users can then curate data with Precisely's Gio AI Assistant and Data Catalog Agent. For example, the agent identifies and tags critical data, flags personal information for oversight and aligns metadata with business processes and regulatory compliance needs.&lt;/p&gt;
 &lt;p&gt;Precisely Data Catalog automatically applies data quality rules and scores to metadata. It also creates data profiles and continuously monitors data health to detect anomalies and other issues. Connectors to more than 20 data sources are currently available, while a variety of others are planned or available by request.&lt;/p&gt;
 &lt;p&gt;The catalog tool also includes the following features:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Use of Precisely's data governance service to add business metadata to catalog entries.&lt;/li&gt; 
  &lt;li&gt;Visualization of data lineage and relationships.&lt;/li&gt; 
  &lt;li&gt;Integration with workflow management and data security tools that are also part of the Precisely platform's foundation.&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;p&gt;&lt;b&gt;Editor's note:&lt;/b&gt;&lt;i&gt; TechTarget editors updated this article in March 2026 for timeliness and to add new information.&lt;/i&gt;&lt;/p&gt;
 &lt;p&gt;&lt;em&gt;George Lawton is a journalist based in London. Over the last 30 years, he has written more than 3,000 stories about computers, communications, knowledge management, business, health and other areas that interest him.&lt;/em&gt;&lt;/p&gt;
 &lt;p&gt;&lt;i&gt;Craig Stedman is an industry editor at TechTarget who creates in-depth packages of content on data technologies and processes.&lt;/i&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>Organizations can use numerous tools to build and manage data catalogs. Here are 15 prominent ones that data leaders should consider for their data management needs.</description>
            <image>https://cdn.ttgtmedia.com/rms/onlineimages/folder-files11.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/feature/16-top-data-catalog-software-tools-to-consider-using</link>
            <pubDate>Mon, 09 Mar 2026 09:00:00 GMT</pubDate>
            <title>15 top data catalog software tools to consider using in 2026</title>
        </item>
        <item>
            <body>&lt;p&gt;Storage governance is increasingly critical as organizations face more data privacy regulations and use analytics to help improve IT operations and business.&lt;/p&gt; 
&lt;p&gt;Governance initiatives had focused on structured data stored in relational databases, but the process has become more complex. Structured,&amp;nbsp;&lt;a href="https://www.techtarget.com/searchstorage/feature/Managing-unstructured-data-to-boost-performance-lower-costs"&gt;unstructured and semistructured data&lt;/a&gt;&amp;nbsp;are all part of storage governance now, so it's crucial for IT leaders to have a comprehensive strategy.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="Why data storage governance is important"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Why data storage governance is important&lt;/h2&gt;
 &lt;p&gt;Organizations are dependent on data, databases, big data, applications and other information resources, so an approved program to protect and manage data resources is essential. Data storage governance provides an understanding of how organizations create, use, authorize, secure, protect, maintain, store, archive and destroy data.&lt;/p&gt;
 &lt;p&gt;&lt;span style="font-size: 16px;"&gt;A data storage governance policy is the first step to achieving a&amp;nbsp;&lt;/span&gt;&lt;a href="https://www.techtarget.com/searchstorage/feature/Explore-secure-data-storage-best-practices" style="font-size: 16px;"&gt;secure storage environment&lt;/a&gt;&lt;span style="font-size: 16px;"&gt;. The policy is the starting point for defining governance procedures and additional policies, such as data storage, cloud storage, data privacy, data protection and data management. These policies are important audit evidence because they establish controls that govern data storage operations.&lt;/span&gt;&lt;/p&gt;
 &lt;figure class="main-article-image full-col" data-img-fullsize="https://www.techtarget.com/rms/onlineImages/data_management-need_to_govern_data-f.png"&gt;
  &lt;img data-src="https://www.techtarget.com/rms/onlineImages/data_management-need_to_govern_data-f_mobile.png" class="lazy" data-srcset="https://www.techtarget.com/rms/onlineImages/data_management-need_to_govern_data-f_mobile.png 960w,https://www.techtarget.com/rms/onlineImages/data_management-need_to_govern_data-f.png 1280w" alt="Why govern data? " height="283" width="560"&gt;
  &lt;figcaption&gt;
   &lt;i class="icon pictures" data-icon="z"&gt;&lt;/i&gt;These are some of the top reasons to have a general data governance program.
  &lt;/figcaption&gt;
  &lt;div class="main-article-image-enlarge"&gt;
   &lt;i class="icon" data-icon="w"&gt;&lt;/i&gt;
  &lt;/div&gt;
 &lt;/figure&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="9 best practices for data storage governance"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;9 best practices for data storage governance&lt;/h2&gt;
 &lt;p&gt;There are several ways organizations can optimize their storage governance, such as with management support and policy training. Best practices include the following:&lt;/p&gt;
 &lt;ol type="1" start="1" class="default-list"&gt; 
  &lt;li&gt;Understand how to establish and&amp;nbsp;&lt;a href="https://www.techtarget.com/searchstorage/Data-storage-management-What-is-it-and-why-is-it-important"&gt;manage storage&lt;/a&gt;, and obtain ongoing support from senior management.&lt;/li&gt; 
  &lt;li&gt;Comprehend the government regulations the organization must follow to&amp;nbsp;&lt;a href="https://www.techtarget.com/searchstorage/tip/6-data-storage-compliance-strategies-for-the-enterprise"&gt;achieve compliance&lt;/a&gt;. Comply with data storage and data management standards. These are audit considerations.&lt;/li&gt; 
  &lt;li&gt;Establish an overall policy for data storage governance, which the organization can supplement with additional policies for data storage management, data privacy, data protection and other issues. The policy can also&amp;nbsp;&lt;a href="https://www.techtarget.com/searchstorage/tip/12-ways-to-manage-your-data-storage-strategy"&gt;define the overall data storage strategy&lt;/a&gt;&amp;nbsp;for the organization. Document procedures for all aspects of data storage. Both policies and procedures are important audit items.&lt;/li&gt; 
  &lt;li&gt;Periodically review and vet data storage governance activities to ensure their relevance and effectiveness. Conduct a periodic&amp;nbsp;&lt;a href="https://www.techtarget.com/searchsecurity/answer/Risk-assessment-vs-risk-analysis-vs-risk-management"&gt;risk analysis&lt;/a&gt;&amp;nbsp;of data storage activities to ensure that the governance program has identified risks, threats and vulnerabilities.&lt;/li&gt; 
  &lt;li&gt;Pay attention to cloud and hybrid storage management. Ensure governance policies cover location, redundancy, access controls and data sovereignty of data stored across cloud, hybrid and multi-cloud environments.&lt;/li&gt; 
  &lt;li&gt;Periodically train and retrain employees who regularly use storage technologies on the proper procedures. New employees should receive training on the proper storage procedures during onboarding.&lt;/li&gt; 
  &lt;li&gt;Establish&amp;nbsp;&lt;a href="https://www.techtarget.com/searchstorage/answer/How-can-organizations-prepare-for-a-data-storage-audit"&gt;IT auditing activities&lt;/a&gt;&amp;nbsp;to ensure the organization follows data storage controls correctly and that those controls are appropriate for the business's requirements. The storage strategy, which ideally is embedded in the storage governance policy, helps define the controls.&lt;/li&gt; 
  &lt;li&gt;Go beyond technology with storage governance. Identify how data storage supports strategic intelligence requirements. Organizations might focus on data value, analytics readiness and compliance, for example.&lt;/li&gt; 
  &lt;li&gt;Regularly communicate with senior IT and company management to reinforce the value of data storage programs and to report on the program's success and how it aligns with goals and objectives.&lt;/li&gt; 
  &lt;li&gt;Establish a team to support the data storage governance program. Members can include technicians, data quality technicians, a chief data officer, and advocates or stewards from key divisions and departments that promote governance activities.&lt;/li&gt; 
 &lt;/ol&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="Overcome challenges to data storage governance"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Overcome challenges to data storage governance&lt;/h2&gt;
 &lt;p&gt;Establish senior management support and a thorough understanding of how data storage helps the firm achieve its goals. These two items greatly help overcome any resistance to a storage governance program.&lt;/p&gt;
 &lt;p&gt;Supplement the above achievements with documented and approved policies and procedures for data storage governance. Once approved, circulate them among all employees so they are aware of the program and comply with the policies.&amp;nbsp;&lt;a target="_blank" href="https://www.lightsondata.com/top-10-data-governance-courses-and-training/" rel="noopener"&gt;Schedule training&lt;/a&gt;&amp;nbsp;on the data storage governance program soon after it launches. Set up refresher training as needed.&lt;/p&gt;
 &lt;p&gt;Periodically review the data storage governance program to keep the program current and aligned with the organization's requirements. Risk assessments help to identify and mitigate any potentially disruptive issues. In addition to program reviews, periodically&amp;nbsp;&lt;a href="https://www.techtarget.com/searchstorage/tip/Perform-data-storage-testing-to-prevent-issues"&gt;test and validate storage procedures&lt;/a&gt;&amp;nbsp;to ensure everything is working.&lt;/p&gt;
&lt;/section&gt;</body>
            <description>Data governance manages the availability, usability, integrity and security of data. Follow these best practices for governance as it relates to data storage.</description>
            <image>https://cdn.ttgtmedia.com/rms/onlineimages/storage_g922017556.jpg</image>
            <link>https://www.techtarget.com/searchstorage/tip/How-to-optimize-data-storage-governance</link>
            <pubDate>Fri, 06 Mar 2026 00:00:00 GMT</pubDate>
            <title>How to optimize data storage governance</title>
        </item>
        <item>
            <body>&lt;p&gt;Technology-led data governance programs are no longer viable. To improve data quality, security, management and stewardship, the entire C-suite must get involved.&lt;/p&gt; 
&lt;p&gt;Many organizations still run data governance through IT, going system by system. In early 2024, executives at professional services firm BDO USA recognized that a holistic, business-led approach to data governance was more effective and created a 45-member &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/Data-governance-roles-and-responsibilities-Whats-needed"&gt;data governance team&lt;/a&gt; of data stewards, business owners and data trustees. According to Mike Gerhard, the chief data and AI officer at BDO, these are executives who provide guidance to achieve company objectives.&lt;/p&gt; 
&lt;p&gt;"We knew we needed data to do a lot of what we wanted to do and to innovate, so we saw the need to transform [the governance program]. We had to change our mindset to see and govern data as a shared commodity," Gerhard said. "Governance today is a team collaboration to make sure we're doing the right thing for the firm and for our clients."&lt;/p&gt; 
&lt;p&gt;To stay competitive, the C-suite must drive data governance and create a culture of shared responsibility in which each function works toward measurable results.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="'Every executive is a data leader'"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;'Every executive is a data leader'&lt;/h2&gt;
 &lt;p&gt;The 2026 Executive Benchmark Survey from technology company &lt;a target="_blank" href="https://www.businesswire.com/news/home/20260203332401/en/Workiva-Executive-Benchmark-Survey-Finds-Instability-is-Accelerating-Data-Automation-and-Governance-in-2026" rel="noopener"&gt;Workiva&lt;/a&gt; found that 79% of business leaders are prioritizing data automation and governance. Moreover, 96% of survey respondents said the CFO, CIO and CSO must unite around a shared &lt;a target="_blank" href="https://www.informationweek.com/data-management/how-to-create-a-sound-data-governance-strategy" rel="noopener"&gt;data governance strategy&lt;/a&gt;, and 96% said better access to shared data improves the likelihood of achieving optimal business outcomes.&lt;/p&gt;
 &lt;p&gt;The results confirm what Gerhard and others are seeing: in a data-driven economy, data's strategic value elevates it to a critical asset that demands attention across the C-suite.&lt;/p&gt;
 &lt;p&gt;"We're now in a world where every executive is a data leader," said Scott Beale, CEO of ISC2, a nonprofit member organization for cybersecurity professionals. "We have to ensure there is shared ownership of data governance and that everyone is aligned on risk, strategic goals and ethics."&lt;/p&gt;
 &lt;p&gt;Beale added that this process is always evolving and improving.&lt;/p&gt;
 &lt;p&gt;"Even organizations that are doing it well could do it better," he said.&lt;/p&gt;
&lt;/section&gt;      
&lt;section class="section main-article-chapter" data-menu-title="How to foster shared responsibility"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;How to foster shared responsibility&lt;/h2&gt;
 &lt;p&gt;Creating an enterprise culture of shared governance within the C-suite isn't easy, experts said. However, organizations can overcome this challenge when the board and the CEO establish the practices and policies that distribute accountability across the leadership team.&lt;/p&gt;
 &lt;p&gt;"It starts with selling a strategy and making sure that strategy is clear," Jon France, CISO of ISC2, said. "It comes with that strategy spark at the top."&lt;/p&gt;
 &lt;p&gt;Beale said chief executives and boards are already doing that work. Leadership is working to &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/7-best-practices-for-successful-data-governance-programs"&gt;mature their data governance programs&lt;/a&gt; because it's essential for automation, analytics and AI.&lt;/p&gt;
 &lt;p&gt;Similarly, Gerhard noted that BDO's business-led collaborative approach provided the firm with more reliable data, which in turn drives more &lt;a href="https://www.techtarget.com/searchitoperations/definition/agentic-process-automation"&gt;automation and AI efforts&lt;/a&gt;. That result has put a brighter spotlight on data governance across the enterprise.&lt;/p&gt;
 &lt;p&gt;"Every CEO now recognizes there is no separating the data strategy and business strategy," Beale said. "They're making sure data is governed in a way that allows the organization to be as competitive as possible and protects trust and enterprise value. They recognize that means data can't live in silos."&lt;/p&gt;
 &lt;p&gt;Beale said that it starts with top enterprise leaders establishing clear rules and a strategy for the C-suite to work together. Leadership must set expectations and determine KPIs so each team member is accountable for performance.&lt;/p&gt;
 &lt;p&gt;The executives might also want to establish a &lt;a href="https://www.techtarget.com/searchcio/definition/steering-committee"&gt;steering committee&lt;/a&gt; to guide the organization, help identify gaps in data ownership and responsibility, and foster an enterprise-wide data governance mindset, said Tom Levi, director, field CISO and cyber strategy at cyber exposure management company CYE. He also advises executives to work through governance scenarios in tabletop exercises to practice a collaborative approach.&lt;/p&gt;
 &lt;p&gt;Gerhard said BDO built its data governance program on the belief that the entire firm owns the data, not any single department. While the firm owns the data, business leaders are accountable for data attributes within their functions, reinforcing a culture of shared responsibility.&lt;/p&gt;
 &lt;p&gt;Along these lines, experts said it's important to note that shared data governance responsibility doesn't make the &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/The-evolution-of-the-chief-data-officer-role"&gt;chief data officer&lt;/a&gt; unnecessary. They stressed that the CDO still oversees data governance even with a team approach.&lt;/p&gt;
 &lt;p&gt;France compared a mature data governance program to financial governance. For example, the CFO is responsible for the organization's financial health, strategy and risk management, even though all executives must be good financial stewards. The same is true for CDOs with data governance.&lt;/p&gt;
&lt;/section&gt;           
&lt;section class="section main-article-chapter" data-menu-title="Challenges to creating shared responsibility"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Challenges to creating shared responsibility&lt;/h2&gt;
 &lt;p&gt;Leadership might find that creating this culture is easier on paper than in practice. The following are common challenges that leadership might encounter:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;&lt;b&gt;Silos.&lt;/b&gt; Sometimes technical or business structures create data silos, which hinder the ability to share responsibility and accountability across the C-suite, Gerhard said. In such cases, organizations might need to restructure these things to facilitate change.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Personalities and motivations&lt;/b&gt;. In other cases, egos and individual priorities can hinder shared responsibility efforts. "Each C-level person brings in their own incentives, priorities and perspectives," Levi said.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Data hoarding.&lt;/b&gt; Board members and executives might not recognize the value of a collaborative governance approach. France said some executives or managers might hoard data. Those business leaders might fear that sharing could affect data use and their ability to succeed in their business objectives.&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Resistance to change. &lt;/b&gt;And in still other cases, an unwillingness to change is a major obstacle. "There may be an executive who has been around for years, and they want to continue to do things their way," Levi said. "Those cultural elements can inhibit the organization from catching up with the market."&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;p&gt;&lt;em&gt;Mary K. Pratt is an award-winning freelance journalist with a focus on covering enterprise IT and cybersecurity management.&lt;/em&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>To improve business outcomes, leadership must move beyond IT controls and adopt a playbook that treats data as a shared enterprise asset with clear roles and policies.</description>
            <image>https://cdn.ttgtmedia.com/rms/onlineimages/collab_a275903017.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/feature/Data-governance-responsibilities-now-belong-in-the-C-suite</link>
            <pubDate>Wed, 04 Mar 2026 13:32:00 GMT</pubDate>
            <title>Data governance responsibilities now belong in the C-suite</title>
        </item>
        <item>
            <body>&lt;p&gt;Enterprises are rapidly adopting AI to unlock its potential, but many fail to address a key area: preparing vast, fragmented data sets so models can use them effectively.&lt;/p&gt; 
&lt;p&gt;AI efforts often slow to a crawl -- or fail entirely -- because teams quickly discover how hard it is to turn raw data into something leaders and systems trust. The real work isn't the modeling but finding the right data, cleaning and governing it, and enforcing standards to keep it consistent and reusable. &lt;a href="https://www.techtarget.com/searchdatamanagement/opinion/2026-will-be-the-year-data-becomes-truly-intelligent"&gt;Enterprises that build momentum&lt;/a&gt; avoid inertia by continuously monitoring, refining and validating their data. Those practices build hard-won trust that the AI needs to produce accurate, relevant results that push projects from experimentation to production.&lt;/p&gt; 
&lt;p&gt;"Prior to the arrival of AI, corporate decision making was centered around the trustworthiness of your existing data, and most people did not [trust their data]," said Stephen Catanzano, an analyst at Omdia, a division of Informa TechTarget.&amp;nbsp;"And our current research shows most people still don't fully trust their data. So, the question remains: can I give my data to an AI agent and have that agent make decisions for my company, like changing processes? Well, you can't. The definition of AI-ready data starts and ends with trust."&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="Why AI data readiness matters now"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Why AI data readiness matters now&lt;/h2&gt;
 &lt;p&gt;This trust factor underscores that AI can't deliver meaningful business value until enterprises address their longstanding gaps in data quality and management.&lt;/p&gt;
 &lt;p&gt;AI data readiness is increasingly recognized as the foundation of successful corporate AI initiatives. Analysts highlight its strategic importance, with&amp;nbsp;Gartner forecasting that 60% of &lt;a href="https://www.techtarget.com/searchcio/feature/AI-failure-examples-What-real-world-breakdowns-teach-CIOs"&gt;AI projects will be abandoned&lt;/a&gt; by the end of 2026 due to inadequate data management. By 2027, the failure rate is expected to climb to 80% for GenAI projects driven by deficiencies in data quality, governance and trust.&lt;/p&gt;
 &lt;p&gt;According to Gartner, siloed data that prevents AI from seeing across multiple CRM, ERP and regulatory systems is a common barrier. Ungoverned data introduces compliance risks and can expose mission-critical data.&lt;/p&gt;
 &lt;p&gt;There's a growing consensus that scalable AI architectures rely on consistent standards to ensure data accuracy, accessibility and compliance. While no single governance model dominates, ISO/IEC 42001 -- the international standard for AI management systems -- offers structured guidance for responsible AI development and oversight. Enterprises often pair it with semantic frameworks, such as Resource Description Framework (RDF) and the Web Ontology Language (OWL). Combined, these approaches strengthen AI data governance, encourage ethical data practices and support scalability.&lt;/p&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="Making data trustworthy takes work"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Making data trustworthy takes work&lt;/h2&gt;
 &lt;p&gt;For many organizations, simply identifying what data they have and where it lives remains the biggest obstacle before they can start the refining process.&lt;/p&gt;
 &lt;p&gt;"It's all well and good to get your data AI ready, but if you don't know where your data resides, it's really hard to do that," said Jack Gold, principal analyst with J. Gold Associates.&amp;nbsp;"Companies have isolated or siloed data stashed all over the place. Unfortunately, a lot of companies are still in the middle of that process. If the majority of users were actually prioritizing this, these data lake companies would be worth trillions."&lt;/p&gt;
 &lt;p&gt;Once organizations know what data they have, governance becomes the next hurdle.&lt;/p&gt;
 &lt;p&gt;"AI systems do not just use data -- they learn from it, and that makes governance critical," Catanzano said. "Poorly governed data leads to biased, insecure, and/or noncompliant AI."&lt;/p&gt;
 &lt;p&gt;Governance gives lineage visibility, allowing teams to trace how the data moves and changes across systems. It also enforces access controls to limit exposure of sensitive information and helps organizations &lt;a href="https://www.techtarget.com/searchenterpriseai/tip/Global-AI-legislation-and-regulation-tracker"&gt;meet regulations, such as HIPAA, GDPR and the EU AI Act&lt;/a&gt;.&lt;/p&gt;
 &lt;p&gt;"Adding in lineage and observability tools is becoming really important," Catanzano said. "They allow you to actually see the data and look for governance challenges, along with being able to map out compliance requirements for data-specific challenges. It has taken people a while to figure out the importance of these tools and how to achieve it. Frankly, most users haven't got to this stage yet."&lt;/p&gt;
&lt;/section&gt;       
&lt;section class="section main-article-chapter" data-menu-title="How embeddings make AI useful"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;How embeddings make AI useful&lt;/h2&gt;
 &lt;p&gt;The next step is making that data relevant at scale. AI performs better when it understands the data's context. Embeddings turn words, images and logs into &lt;a href="https://www.techtarget.com/whatis/definition/vector"&gt;vectors&lt;/a&gt;, so systems retrieve the right content rather than guessing.&lt;/p&gt;
 &lt;p&gt;Paired with strong metadata, embeddings help AI return the right information at the right time. AI is shifting toward metadata-rich, vector-based retrieval, Catanzano said.&lt;/p&gt;
 &lt;p&gt;"We have been moving towards higher levels of metadata and larger amounts of the vectorization of data," Catanzano noted. "So now, users want to use more AI because vectors create relevancy, which means AI can find the most relevant data based on vector scores and so improve the quality of data being searched for."&lt;/p&gt;
 &lt;p&gt;Vectorization is becoming a baseline requirement for effective AI implementation, Catanzano noted. Converting unstructured data into embeddings and combining them with high-fidelity metadata optimizes retrieval precision and &lt;a href="https://www.techtarget.com/searchdatamanagement/opinion/Data-intelligence-isnt-just-a-buzzword"&gt;increases confidence in the accuracy&lt;/a&gt; of the information delivered to the end user.&lt;u&gt; &lt;/u&gt;&lt;/p&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="How tokenization improves performance"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;How tokenization improves performance&lt;/h2&gt;
 &lt;p&gt;With retrieval grounded in embeddings and metadata, the next lever on performance is how text is prepared for the model. Using enterprise data effectively with AI workflows often requires converting it into formats the language models can understand.&lt;/p&gt;
 &lt;p&gt;&lt;a href="https://www.techtarget.com/searchsecurity/definition/tokenization"&gt;Tokenization&lt;/a&gt; is a key part of the pipeline, but it's only the first step. Once tokenized, the model applies learned patterns and relationships to analyze content, generate responses or make predictions. &lt;a href="https://www.techtarget.com/searchenterpriseai/feature/Can-tokenization-free-up-more-data-for-AI-model-training"&gt;Efficient tokenization&lt;/a&gt; reduces the number of tokens the system must handle, improving response times and lowering compute and inference costs in the production environment.&lt;/p&gt;
 &lt;p&gt;In modern AI workflows, organizations often convert documents and other unstructured data into &lt;a href="https://www.techtarget.com/searchenterpriseai/definition/vector-embeddings"&gt;vector embeddings&lt;/a&gt;. This change makes the information available for wider use, enabling more precise insights tied to specific business needs.&lt;/p&gt;
 &lt;p&gt;"Developers and users have to transform their data located, for instance, in a database, into a [format] that can travel across platforms," said Frank Dzubeck, president of Communications Network Architects. "Companies, in the pharmaceutical industry, for instance, are doing that now. It changes the way they can look at data because they can create [embeddings] that specifically address problems they are researching in their industry."&lt;/p&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="What are the building blocks of AI‑ready data?"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;What are the building blocks of AI‑ready data?&lt;/h2&gt;
 &lt;p&gt;Beyond AI data governance requirements, enterprises achieve the best results when embeddings rest on a strong data layer consisting of several key elements.&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;&lt;b&gt;Standardized structures. &lt;/b&gt;Common formats, such as CSV and JSON, help keep data portable, but their real value is applying consistency to the information they hold;&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Smart labeling.&lt;/b&gt; Tagging and annotating data ensure AI models can interpret the raw values and the intended meaning behind them;&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;A shared language.&lt;/b&gt; Semantic frameworks, such as RDF and Shapes Constraint Language (SHACL), act as a translator to give data sets a common structure to promote interoperability; and&lt;/li&gt; 
  &lt;li&gt;&lt;b&gt;Deep context. &lt;/b&gt;Using logic tools, such as OWL, gives the context and meaning needed to form relationships across data sets.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;   
&lt;section class="section main-article-chapter" data-menu-title="What to ask vendors before you buy"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;What to ask vendors before you buy&lt;/h2&gt;
 &lt;p&gt;Don't be swayed by a polished demo. Press for specifics about how the system handles your data today and what could break tomorrow.&lt;/p&gt;
 &lt;p&gt;&lt;a href="https://www.techtarget.com/searchdatamanagement/opinion/How-data-for-AI-is-changing-the-modern-data-platform"&gt;When evaluating vendors&lt;/a&gt;, anchor your questions in your governance model and data flows. Start with data use and ownership. How is proprietary data isolated? Will your content be used to train shared models, or will it only serve your tenants? What are the defaults for retention, deletion and cross-tenant safeguards?&lt;/p&gt;
 &lt;p&gt;Models trained on your organization's data tend to generate more accurate results that are less susceptible to biases inherent in internet-based data. Ask how the tuning is done and how quality is measured over time.&lt;/p&gt;
 &lt;p&gt;It's also critical to understand the data preparation workflow and the tools involved. What formats are supported? How is the data transformed? What lineage, logging and rollback is available? What are the commitments for backward compatibility as the platform evolves?&lt;/p&gt;
 &lt;p&gt;Without these answers, organizations &lt;a target="_blank" href="https://edmcouncil.org/frameworks/cdmc/14-key-controls/" rel="noopener"&gt;risk&lt;/a&gt; getting locked into proprietary formats that might not work in later versions.&lt;/p&gt;
 &lt;p&gt;"They need to actually show users their various transformation tools for databases, searching and other functions because they are all different," said Dzubeck. "As for future proofing, that's a tough question for users to get an answer to. They are only going to give you answers that directly link to their products and strategies like Google, which is focused on search, and large database makers like Oracle."&lt;/p&gt;
 &lt;p&gt;&lt;i&gt;Ed Scannell is a freelance writer and journalist based in Needham, Mass. He reports on a wide range of technologies and issues related to corporate IT. He can be reached at ed.scannell@gmail.com.&lt;/i&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>As organizations dive into AI adoption, many realize the first real bottleneck is not the model but how to prepare their information so it can be used effectively in AI workflows.</description>
            <image>https://cdn.ttgtmedia.com/rms/onlineimages/ai_a352095729.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/feature/AI-data-governance-guidance-that-gets-you-to-the-finish-line</link>
            <pubDate>Tue, 03 Mar 2026 07:36:00 GMT</pubDate>
            <title>AI data governance guidance that gets you to the finish line</title>
        </item>
        <item>
            <body>&lt;p&gt;Big data environments in organizations are only getting bigger. The ever-increasing volume and variety of data collected in them requires investments in big data tools to support analytics and AI applications. But choosing the right technologies is complicated: Enterprise data leaders have a wide variety of tools to consider.&lt;/p&gt; 
&lt;p&gt;The available choices include numerous open source big data tools, many of which are offered by technology vendors in commercial versions or as part of big data platforms. The following are 18 popular open source technologies for &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/How-to-build-an-enterprise-big-data-strategy-in-4-steps"&gt;managing and analyzing big data&lt;/a&gt;, listed in alphabetical order with an overview of each one's features, capabilities and potential uses. TechTarget editors compiled the list based on their research of available technologies and analysis from consulting firms such as Forrester Research and Gartner.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="1. Airflow"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;1. Airflow&lt;/h2&gt;
 &lt;p&gt;Apache Airflow is a workflow management platform for scheduling and running &lt;a href="https://www.techtarget.com/searchbusinessanalytics/news/365534255/Data-pipelines-deliver-the-fuel-for-data-science-analytics"&gt;complex data pipelines&lt;/a&gt; in big data systems. It enables data engineers and other users to ensure each task in a workflow can access the required system resources and is executed in the designated order. Airflow is most commonly used to orchestrate data integration and transformation processes, machine learning (ML) operations, business applications and IT infrastructure management tasks, but it also supports other types of workflows.&lt;/p&gt;
 &lt;p&gt;The platform has a modular architecture built around directed acyclic graphs that illustrate the dependencies between workflow tasks. Airflow pipelines are defined in Python and can be generated dynamically. Airbnb initially created Airflow for internal use, and the technology became a top-level project within the Apache Software Foundation in 2019.&lt;/p&gt;
 &lt;p&gt;Airflow also includes the following features:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Time- and dependency-based scheduling of workflows, plus an event-driven scheduling option.&lt;/li&gt; 
  &lt;li&gt;A web application UI to visualize data pipelines, monitor their production status and troubleshoot problems.&lt;/li&gt; 
  &lt;li&gt;Ready-made integrations with major cloud platforms and other third-party services.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="2. Delta Lake"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;2. Delta Lake&lt;/h2&gt;
 &lt;p&gt;Delta Lake is a table storage layer that can be used to &lt;a href="https://www.techtarget.com/searchdatamanagement/news/366545117/Lakehouse-architecture-the-best-fit-for-modern-data-needs"&gt;build a data lakehouse architecture&lt;/a&gt; combining elements of data lakes and data warehouses. The Delta Lake framework creates a unified format for structured, semistructured and unstructured data, eliminating data silos that often &lt;a href="https://www.techtarget.com/searchdatamanagement/tip/10-big-data-challenges-and-how-to-address-them"&gt;stymie big data applications&lt;/a&gt;. It also provides common semantics for both batch and stream processing of table reads and writes.&lt;/p&gt;
 &lt;p&gt;To ensure data integrity, Delta Lake supports transactions that adhere to the four &lt;a href="https://www.techtarget.com/searchdatamanagement/definition/ACID"&gt;ACID properties&lt;/a&gt;: atomicity, consistency, isolation and durability. A liquid clustering capability optimizes how data is stored based on query patterns, offering an alternative to traditional data partitioning. Databricks, a software vendor founded by the creators of the Apache Spark processing engine, developed Delta Lake and made the Spark-compatible technology open source in 2019 through the Linux Foundation.&lt;/p&gt;
 &lt;p&gt;Delta Lake also includes the following features:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Support for storing data in an open Apache Parquet format.&lt;/li&gt; 
  &lt;li&gt;Delta Universal Format, a function commonly known as UniForm that enables Delta Lake tables to be read in Iceberg and Hudi, two other Parquet-based table formats.&lt;/li&gt; 
  &lt;li&gt;A time-travel capability that provides access to earlier versions of data sets for audits and rollbacks.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="3. Drill"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;3. Drill&lt;/h2&gt;
 &lt;p&gt;Apache Drill is a low-latency distributed query engine best suited for workloads involving large, complex data sets with diverse types of records and fields. The Drill website claims it can scale across thousands of cluster nodes and query petabytes of data using SQL and standard connectivity APIs. It handles a combination of structured and semistructured data, including nested data types such as JSON and Parquet files.&lt;/p&gt;
 &lt;p&gt;Drill is built on a schema-free JSON document model and layers on top of multiple data sources, enabling users to query a wide range of data in different formats. It supports various file types and sources, including Hadoop SequenceFiles and event logs, &lt;a href="https://www.techtarget.com/searchcloudcomputing/tip/Compare-NoSQL-database-types-in-the-cloud"&gt;NoSQL databases&lt;/a&gt; and cloud object storage. Drill users can store multiple files in a directory and query them as a single entity.&lt;/p&gt;
 &lt;p&gt;First released in 2015, the software can also do the following:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Query data in most relational databases through a plugin.&lt;/li&gt; 
  &lt;li&gt;Work with &lt;a href="https://www.techtarget.com/searchbusinessanalytics/feature/Top-business-intelligence-tools-to-know-about"&gt;commonly used BI tools&lt;/a&gt;, such as Tableau and Qlik Sense.&lt;/li&gt; 
  &lt;li&gt;Run in any distributed cluster environment, although Apache ZooKeeper must be installed along with it to maintain information about cluster configurations.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="4. Druid"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;4. Druid&lt;/h2&gt;
 &lt;p&gt;Apache Druid is a real-time analytics database with an interactive query engine that provides low query latency, high user concurrency, multi-tenant capabilities and instant visibility into streaming data. Hundreds or thousands of end users can simultaneously query data stored in Druid with no effect on performance, according to its developers.&lt;/p&gt;
 &lt;p&gt;Written in Java and created in 2011, Druid became an Apache technology in 2018. Best suited for storing event-driven data, it's considered a high-performance alternative to traditional data warehouses. Like a data warehouse, Druid uses column-oriented storage and can load files in batch mode. However, it also incorporates features from search systems and time series databases, including the following:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Compressed bitmap indexes to speed up searches and data filtering.&lt;/li&gt; 
  &lt;li&gt;Time-based data partitioning and querying.&lt;/li&gt; 
  &lt;li&gt;Flexible schemas with native support for semistructured data and nested data structures.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="5. Flink"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;5. Flink&lt;/h2&gt;
 &lt;p&gt;Another Apache technology, Flink is a &lt;a href="https://www.techtarget.com/searchdatamanagement/definition/stream-processing"&gt;stream processing&lt;/a&gt; framework for high-performance distributed applications, including always-available ones. It supports stateful computations over both bounded and unbounded data streams and can be used for batch, graph and iterative processing. One of the main benefits touted by Flink's proponents is its speed: The software processes millions of events in real time with low latency and high throughput.&lt;/p&gt;
 &lt;p&gt;Flink began as a university research initiative in Germany and became an Apache project in 2014. In addition to event-driven applications -- such as fraud or anomaly detection -- potential use cases include continuous data pipelines and both streaming and batch analytics. Flink runs in all common cluster environments and also includes the following features:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;In-memory computations with the ability to access disk storage when needed.&lt;/li&gt; 
  &lt;li&gt;Three layers of APIs for creating different types of applications.&lt;/li&gt; 
  &lt;li&gt;A set of libraries for complex event processing, ML and other &lt;a href="https://www.techtarget.com/searchbusinessanalytics/feature/8-big-data-use-cases-for-businesses-and-industry-examples"&gt;common big data use cases&lt;/a&gt;.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="6. Hadoop"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;6. Hadoop&lt;/h2&gt;
 &lt;p&gt;Apache Hadoop is a distributed framework for storing data and running applications on commodity hardware clusters. First released in 2006 as a pioneering big data technology, it helps users handle large volumes of structured, unstructured and semistructured data. Hadoop is also at the center of a broader technology ecosystem that includes various related tools and frameworks for processing, managing and analyzing big data. While Hadoop has been partially eclipsed by Spark and other technologies, it's still used by many organizations.&lt;/p&gt;
 &lt;p&gt;Hadoop includes these primary components:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;The Hadoop Distributed File System (&lt;a href="https://www.techtarget.com/searchdatamanagement/definition/Hadoop-Distributed-File-System-HDFS"&gt;HDFS&lt;/a&gt;) splits data into blocks for storage on cluster nodes, uses replication methods to prevent data loss and manages access to the data.&lt;/li&gt; 
  &lt;li&gt;Hadoop YARN schedules data processing jobs to run on cluster nodes and allocates system resources to them.&lt;/li&gt; 
  &lt;li&gt;Hadoop MapReduce, a built-in batch processing engine, splits up large computations and runs them on different nodes for speed and load balancing.&lt;/li&gt; 
  &lt;li&gt;Hadoop Common is a shared set of utilities and libraries.&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;p&gt;Initially, Hadoop was limited to running MapReduce batch applications. The addition of YARN in 2013 opened it up to other processing engines and use cases, but the framework is still most commonly used with MapReduce.&lt;/p&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="7. Hive"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;7. Hive&lt;/h2&gt;
 &lt;p&gt;Also an Apache technology, Hive is SQL-based &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Evaluating-your-need-for-a-data-warehouse-platform"&gt;data warehouse infrastructure software&lt;/a&gt; for reading, writing and managing large data sets in distributed Hadoop storage environments. It runs on top of Hadoop and processes structured data for summarization, querying and analysis. Hive supports ACID transactions, low-latency analytical processing and cost-based query optimization, the latter through integration with the Apache Calcite tool.&lt;/p&gt;
 &lt;p&gt;In addition to HDFS files, Hive can access ones stored in the Apache HBase database and other systems. It also enables users to create and read Iceberg tables. Hive Metastore Server, its central metadata repository, provides data abstraction and data discovery features similar to those in traditional data warehouses. Facebook created Hive for internal use, and it became an Apache top-level project in 2010.&lt;/p&gt;
 &lt;p&gt;Other key features include the following:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;HiveQL, a language with standard SQL functionality for data querying and analytics.&lt;/li&gt; 
  &lt;li&gt;Native support for cloud object storage services.&lt;/li&gt; 
  &lt;li&gt;MapReduce, Spark and Apache Tez as execution back-end options.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="8. HPCC Systems"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;8. HPCC Systems&lt;/h2&gt;
 &lt;p&gt;HPCC Systems is a big data processing platform that LexisNexis Risk Solutions developed as an alternative to Hadoop and Spark. Befitting its full name -- High-Performance Computing Cluster Systems -- the technology supports data-intensive applications requiring speed and scalability on clusters built from commodity hardware. Its primary use case is enabling rapid data engineering for analytics applications in data lake environments.&lt;/p&gt;
 &lt;p&gt;The platform includes these main components:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Thor, a data refinery engine used to cleanse, merge and transform data for use in queries.&lt;/li&gt; 
  &lt;li&gt;Roxie, a data delivery engine that serves prepared data from the refinery to end users for querying.&lt;/li&gt; 
  &lt;li&gt;Enterprise Control Language, a programming language commonly known as ECL that's used for data management and query processing.&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;p&gt;HPCC Systems also includes a library of ML algorithms, plus tools for monitoring clusters and profiling, curating and governing data. While still primarily overseen by LexisNexis, it became open source in 2011 and is freely available to download under the Apache 2.0 license. The current release is a cloud-native platform that runs in Docker containers on &lt;a href="https://www.techtarget.com/searchitoperations/definition/Google-Kubernetes"&gt;Kubernetes&lt;/a&gt; in both the AWS and Microsoft Azure clouds. Deployments of the original bare-metal platform are also still supported.&lt;/p&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="9. Hudi"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;9. Hudi&lt;/h2&gt;
 &lt;p&gt;Apache Hudi -- pronounced &lt;i&gt;hoodie&lt;/i&gt; -- is a platform for managing large analytics data sets stored in HDFS and other Hadoop-compatible file systems deployed in cloud object storage services. Short for "Hadoop upserts, deletes and incrementals," Hudi provides database-like functionality for ingesting and updating data to support real-time analytics in data lakes and lakehouses.&lt;/p&gt;
 &lt;p&gt;First developed by Uber and an Apache top-level project since 2020, Hudi is built on an open table format that supports both Parquet and Apache ORC as the base file format. The platform integrates with Spark, Flink and other data processing and query engines. It supports ACID transactions, multimodal indexing to boost query performance and historical data analysis through a time-travel feature.&lt;/p&gt;
 &lt;p&gt;Hudi also includes a data management framework that organizations can use to do the following:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Simplify incremental data processing and data pipeline development.&lt;/li&gt; 
  &lt;li&gt;Improve &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Data-quality-for-big-data-Why-its-a-must-and-how-to-improve-it"&gt;data quality in big data systems&lt;/a&gt;.&lt;/li&gt; 
  &lt;li&gt;Manage data set lifecycles.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="10. Iceberg"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;10. Iceberg&lt;/h2&gt;
 &lt;p&gt;Another Apache technology, Iceberg is an open table format for managing large analytics data sets &lt;a href="https://www.techtarget.com/searchdatamanagement/opinion/Why-Apache-Iceberg-is-essential-for-modern-data-lakehouses"&gt;stored in data lakes and lakehouses&lt;/a&gt;. According to the project's website, Iceberg is typically used in applications where individual tables contain tens of petabytes of data. The tables can be read from a single cluster node, without requiring a distributed SQL engine to sort through metadata and find the files needed for queries.&lt;/p&gt;
 &lt;p&gt;To boost query performance, Iceberg tracks individual data files in tables rather than directories, using metadata files to maintain a snapshot log of changes to a table. It supports SQL commands to update, merge or delete data and enables multiple query engines to simultaneously read and write data in a single table. Created by Netflix for internal use, Iceberg became an Apache top-level project in 2020.&lt;/p&gt;
 &lt;p&gt;Other notable features include the following:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Schema evolution for modifying tables without rewriting or migrating data.&lt;/li&gt; 
  &lt;li&gt;Hidden partitioning that frees users from maintaining partitions and automatically updates table layouts as data or queries change.&lt;/li&gt; 
  &lt;li&gt;A time-travel capability, plus version rollback for resetting tables to a known good state.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="11. Kafka"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;11. Kafka&lt;/h2&gt;
 &lt;p&gt;Apache Kafka is a distributed event streaming platform that supports data pipelines, data integration, streaming analytics and critical business applications. Created by LinkedIn and handed over to Apache in 2011, Kafka handles petabytes of data and trillions of event messages per day. It uses a publish-subscribe model to transmit messages and enables users to store event streams in distributed, fault-tolerant clusters for long-term use. Data streams can similarly be processed on the fly or later.&lt;/p&gt;
 &lt;p&gt;To boost scalability, Kafka decouples applications that produce and consume event data and partitions the data across multiple storage servers, which are called &lt;i&gt;brokers&lt;/i&gt;. It can be deployed on bare-metal hardware or in VMs and containers, both on-premises and in the cloud.&lt;/p&gt;
 &lt;p&gt;The following are some of Kafka's other key components:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;A set of six core APIs for Java and the Scala programming language.&lt;/li&gt; 
  &lt;li&gt;Built-in stream processing capabilities for joining, aggregating, filtering and transforming data.&lt;/li&gt; 
  &lt;li&gt;Elastic scalability to up to 1,000 brokers per cluster.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="12. Kylin"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;12. Kylin&lt;/h2&gt;
 &lt;p&gt;Apache Kylin is a distributed data warehouse and online analytical processing (&lt;a href="https://www.techtarget.com/searchdatamanagement/definition/OLAP"&gt;OLAP&lt;/a&gt;) platform designed to support large data sets and queries involving trillions of records. Kylin's storage layer is built on top of Delta Lake and Parquet. The platform includes a native compute engine added in 2024 that's based on Spark and Apache Gluten, a performance accelerator plugin for Spark.&lt;/p&gt;
 &lt;p&gt;Internal data tables that Kylin manages directly were added along with the native engine. Kylin also still supports tables imported from data sources such as Hive, Kafka and Iceberg, but the internal tables offer greater flexibility for querying data. It provides a SQL interface for querying data and connects to Excel and BI tools such as Tableau and Microsoft Power BI. Initially developed by eBay, Kylin became an Apache top-level project in 2015.&lt;/p&gt;
 &lt;p&gt;Kylin also offers the following features:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Precalculation of multidimensional OLAP cubes to improve query performance.&lt;/li&gt; 
  &lt;li&gt;A data modeling and indexing recommendation engine.&lt;/li&gt; 
  &lt;li&gt;Combined analysis of streaming and batch data.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="13. Pinot"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;13. Pinot&lt;/h2&gt;
 &lt;p&gt;Also an Apache project, Pinot is a real-time distributed OLAP data store that supports low-latency querying in analytics applications. According to its developers, Pinot handles petabytes of data containing trillions of records and concurrently processes hundreds of thousands of queries per second. To deliver the promised performance, Pinot has a fault-tolerant architecture with no single point of failure and supports horizontal scaling of clusters. Other configuration changes can also be done dynamically without affecting data availability or query performance.&lt;/p&gt;
 &lt;p&gt;Pinot uses a columnar storage format and offers various indexing techniques to filter, aggregate and group data. To simplify data storage and replication, the system assumes all stored data is immutable. However, it supports upserts to keep streaming data sets up to date, as well as background purges of sensitive data to comply with privacy laws. Created by LinkedIn for internal use, Pinot became an Apache top-level project in 2021.&lt;/p&gt;
 &lt;p&gt;The following features are also included:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Near-real-time data ingestion from streaming sources, plus batch ingestion from HDFS, Spark and cloud storage services.&lt;/li&gt; 
  &lt;li&gt;A SQL interface for interactive querying and a REST API for programming queries.&lt;/li&gt; 
  &lt;li&gt;Integration with ZooKeeper for distributed metadata storage and Apache Helix for cluster management.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="14. Presto"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;14. Presto&lt;/h2&gt;
 &lt;p&gt;Presto is a SQL query engine optimized for low-latency querying of large data sets. It supports analytics applications across multiple petabytes of data in data lakes, data lakehouses and other repositories. To further boost performance and reliability, Presto's developers are converting its core execution engine from Java to a C++ version based on Velox, an open source acceleration library. An early version of Presto C++ is available, but it has a limited set of connectors and doesn't support some of Presto's built-in query functions.&lt;/p&gt;
 &lt;p&gt;Presto's development began at Facebook. When its creators left the company in 2018, the technology split into two branches: PrestoDB, which Facebook still led, and PrestoSQL, led by the original developers. In 2020, PrestoDB reverted to Presto, and PrestoSQL was renamed Trino. The Presto open source project is now overseen by the Presto Foundation, which is part of the Linux Foundation.&lt;/p&gt;
 &lt;p&gt;Presto also includes the following features:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Connectors to 36 data sources, including Delta Lake, Druid, Hive, Hudi, Iceberg, Pinot and various databases.&lt;/li&gt; 
  &lt;li&gt;The ability to combine data from multiple sources in a single query.&lt;/li&gt; 
  &lt;li&gt;A web-based UI and a CLI for querying, plus support for the Apache Superset data exploration tool.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="15. Samza"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;15. Samza&lt;/h2&gt;
 &lt;p&gt;Apache Samza is a distributed stream processing system that enables users to build stateful applications for real-time processing of data from Kafka, HDFS and several other sources. It then writes the processed data back to some of the same systems. Use cases for Samza include event-based applications, real-time analytics and extract, transform and load (ETL) processes on streaming data.&lt;/p&gt;
 &lt;p&gt;The Samza website says it can handle "several terabytes" of state data, with low latency and high throughput for data analysis. The system also supports stateless stream processing. It runs on top of Hadoop YARN or in a standalone deployment mode; the latter option enables Samza to be a component of larger applications and lets users implement Kubernetes or another cluster manager instead of YARN. Originally developed by LinkedIn, Samza has been an Apache top-level project since 2015.&lt;/p&gt;
 &lt;p&gt;Other features include the following:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;A pair of high- and low-level APIs for different use cases, plus a declarative SQL interface.&lt;/li&gt; 
  &lt;li&gt;The ability to run as a lightweight embedded library in Java and Scala applications.&lt;/li&gt; 
  &lt;li&gt;Fault-tolerant features for migrating tasks in the event of system failures and rapidly recovering from them.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="16. Spark"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;16. Spark&lt;/h2&gt;
 &lt;p&gt;Apache Spark is a unified data processing and analytics engine used for data engineering in both batch and streaming applications, as well as for interactive querying, ML and exploratory data analysis. Spark often outperforms MapReduce on batch processing, making it the top choice for such tasks in many &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Building-a-big-data-architecture-Core-components-best-practices"&gt;big data environments&lt;/a&gt;. It's also widely used as a large-scale analytics platform.&lt;/p&gt;
 &lt;p&gt;Spark includes the following core modules and libraries to support its various use cases:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Spark SQL, for processing structured and unstructured data via SQL queries.&lt;/li&gt; 
  &lt;li&gt;Spark Structured Streaming, a module for building streaming applications and data pipelines.&lt;/li&gt; 
  &lt;li&gt;MLlib, a machine learning library that includes various algorithms and related utilities.&lt;/li&gt; 
  &lt;li&gt;Dataset and DataFrame APIs, which are used to organize distributed data sets for processing.&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;p&gt;Spark runs on clusters managed by Hadoop YARN, Kubernetes or a standalone clustering tool built into the platform. It handles data from various sources, including HDFS, flat files and both relational and NoSQL databases. In addition to SQL, Spark supports Python, Scala, Java and R for programming. A Spark Connect feature enables client applications to connect to remote servers, simplifying development and deployment. Spark was created at the University of California, Berkeley, in 2009 and became an Apache top-level project in 2014.&lt;/p&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="17. Storm"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;17. Storm&lt;/h2&gt;
 &lt;p&gt;Storm, another Apache technology, is a distributed real-time computation system for processing unbounded data streams. Its use cases include real-time analytics, ML, continuous computation and ETL procedures on streaming data. The fault-tolerant system guarantees data processing, with multiple guarantee levels available to meet different application needs.&lt;/p&gt;
 &lt;p&gt;The Apache Storm website says it can integrate with any message queueing system or database to access streaming data. Storm also supports any programming language for application development, and the system's out-of-the-box cluster configurations are suitable for production use. ZooKeeper is integrated to coordinate Storm clusters.&lt;/p&gt;
 &lt;p&gt;Storm became an Apache top-level project in 2014 and also includes the following elements:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;A basic API and Trident, a higher-level interface for processing data in Storm&lt;/li&gt; 
  &lt;li&gt;Inherent parallelism that supports high data throughput with low latency.&lt;/li&gt; 
  &lt;li&gt;An experimental Storm SQL feature that enables SQL queries to run against streaming data sets.&lt;/li&gt; 
 &lt;/ul&gt;
&lt;/section&gt;     
&lt;section class="section main-article-chapter" data-menu-title="18. Trino"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;18. Trino&lt;/h2&gt;
 &lt;p&gt;As mentioned above, Trino branched off from the Presto query engine and was originally named PrestoSQL. Like Presto, it's a distributed SQL engine for use in big data analytics applications. According to the Trino website, it supports low-latency analytics in exabyte-scale data lakes and lakehouses, as well as large data warehouses.&lt;/p&gt;
 &lt;p&gt;Trino includes built-in connectors to 25 data sources, and seven external connectors are also available. It provides an interactive CLI for querying data, plus a plugin that lets users run queries in Grafana, an open source data visualization and dashboard design tool. In addition, Trino works with Tableau, Power BI and other BI and analytics tools, as well as Apache Superset and R.&lt;/p&gt;
 &lt;p&gt;Trino is overseen by the Trino Software Foundation and also supports the following capabilities:&lt;/p&gt;
 &lt;ul class="default-list"&gt; 
  &lt;li&gt;Both ad hoc interactive analytics and long-running batch queries.&lt;/li&gt; 
  &lt;li&gt;Queries that combine data from multiple sources through a federation feature.&lt;/li&gt; 
  &lt;li&gt;Deployment in Kubernetes clusters and Docker containers.&lt;/li&gt; 
 &lt;/ul&gt;
 &lt;p&gt;&lt;b&gt;Editor's note: &lt;/b&gt;&lt;i&gt;TechTarget editors updated this article in February 2026 for timeliness and to add new information.&lt;/i&gt;&lt;/p&gt;
 &lt;p&gt;&lt;i&gt;Mary K. Pratt is an award-winning freelance journalist with a focus on covering enterprise IT and cybersecurity management.&lt;/i&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>Numerous tools are available to use in big data applications. Here are 18 popular open source big data technologies, with details on their key features and use cases.</description>
            <image>https://cdn.ttgtmedia.com/visuals/searchDataManagement/integration_technology/datamanagement_article_015.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/feature/15-big-data-tools-and-technologies-to-know-about</link>
            <pubDate>Thu, 26 Feb 2026 00:00:00 GMT</pubDate>
            <title>18 top big data tools and technologies to know about in 2026</title>
        </item>
        <item>
            <body>&lt;p&gt;When it comes to acting on data, timing is everything.&lt;/p&gt; 
&lt;p&gt;What organizations once analyzed yesterday, they now need to understand immediately. Real-time data streaming is becoming essential infrastructure for competitive AI applications, and the gap between companies that can use it and those that can't is widening.&lt;/p&gt; 
&lt;section class="section main-article-chapter" data-menu-title="Why real-time matters now"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Why real-time matters now&lt;/h2&gt;
 &lt;p&gt;The shift from batch processing to real-time streaming is more than a technical upgrade -- it's a fundamental change in how businesses operate. Traditional approaches to data analysis, where information is collected throughout the day and processed in scheduled batches, made sense when business moved more slowly. That world is disappearing.&lt;/p&gt;
 &lt;p&gt;Markets move in milliseconds. Customers expect instant personalization. Operational issues must be caught before they cascade into failures.&lt;/p&gt;
 &lt;p&gt;Consider fraud detection in financial services. Identifying a suspicious transaction in real time can prevent the crime, while discovering it hours later during a batch review usually means investigating after the fact. In manufacturing, streaming sensor data from equipment &lt;a href="https://www.techtarget.com/searchbusinessanalytics/feature/Real-time-edge-analytics-use-cases-for-business"&gt;enables teams to predict failures proactively&lt;/a&gt;, not just analyze why something broke. In retail, live analysis of browsing behavior and inventory levels enables dynamic pricing and personalization that batch processing cannot deliver.&lt;/p&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="Implementing streaming with guardrails"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Implementing streaming with guardrails&lt;/h2&gt;
 &lt;p&gt;The challenge is that the data volume and velocity have exploded. Connected devices, digital transactions and user interactions generate continuous streams of information that require immediate analysis to create value. The tools and infrastructure exist, but implementing them effectively requires rethinking how organizations approach data strategy.&lt;/p&gt;
 &lt;p&gt;Real-time streaming introduces a layer of complexity that many enterprises underestimate. It's not just about speed. It's about building systems that handle continuous data flows, &lt;a href="https://www.techtarget.com/searchenterpriseai/feature/Optimize-AI-models-to-generate-more-bang-for-your-buck"&gt;integrate with AI models&lt;/a&gt; that make instant decisions and remain reliable when delays occur or failures ripple through the pipeline. The technical demands are significant, and the margin for error is small.&lt;/p&gt;
 &lt;p&gt;Data governance is also more complicated in real-time environments. All the concerns about data quality, privacy and compliance that exist in batch systems become more acute when data flows continuously. Organizations need to implement controls that ensure compliance and data integrity without increasing latency.&lt;/p&gt;
&lt;/section&gt;    
&lt;section class="section main-article-chapter" data-menu-title="Start where seconds make a difference"&gt;
 &lt;h2 class="section-title"&gt;&lt;i class="icon" data-icon="1"&gt;&lt;/i&gt;Start where seconds make a difference&lt;/h2&gt;
 &lt;p&gt;Despite these challenges, the competitive advantages are compelling. Enterprises that successfully implement real-time AI capabilities can respond to market changes faster than competitors, deliver superior customer experiences and improve operations to create measurable business value. The question isn't whether to embrace this shift, but how quickly you can do so effectively.&lt;/p&gt;
 &lt;p&gt;Start by identifying where instantaneous analysis delivers the most value. Not every use case requires real-time data. Some assessments are &lt;a href="https://www.techtarget.com/searchdatamanagement/feature/Building-a-big-data-architecture-Core-components-best-practices"&gt;perfectly suited to batch processing.&lt;/a&gt; The goal is to focus on applications where immediate insights drive timely action: fraud prevention, dynamic optimization, real-time personalization or operational monitoring where minutes matter.&lt;/p&gt;
 &lt;p&gt;Infrastructure decisions must align with business objectives. Stakeholders should evaluate streaming platforms based on throughput requirements, latency tolerances, integration with existing systems and operational complexity. Cloud-based products offer ease of deployment but may introduce vendor lock-in. Open-source options provide flexibility but require more internal expertise.&lt;/p&gt;
 &lt;p&gt;Integration with AI systems is critical. Models should be optimized for low-latency inference. &lt;a href="https://www.techtarget.com/searchenterpriseai/feature/How-to-build-a-machine-learning-model-in-7-steps"&gt;Feature engineering pipelines&lt;/a&gt; must support both batch and streaming data. The entire system should be monitored to catch issues before they affect business outcomes.&lt;/p&gt;
 &lt;p&gt;Most importantly, organizations need to build the cultural capability to work with real-time data. This requires cross-functional teams that understand both the business context and the technical requirements, workflows that enable rapid experimentation, and a willingness to iterate and improve as operational needs evolve.&lt;/p&gt;
 &lt;p&gt;Real-time data streaming isn't a future capability -- it's a basic &lt;a target="_blank" href="https://hbr.org/2026/02/why-your-digital-investments-arent-creating-value" rel="noopener"&gt;expectation&lt;/a&gt;. Organizations that recognize this and invest accordingly will define the competitive landscape in their industries. Those that don't will find themselves perpetually reacting to competitors who can see and act on opportunities they're still processing.&lt;/p&gt;
 &lt;p&gt;&lt;i&gt;Stephen Catanzano is a senior analyst at Omdia where he covers data management and analytics.&lt;/i&gt;&lt;/p&gt;
 &lt;p&gt;&lt;i&gt;Omdia is a division of&amp;nbsp;Informa TechTarget.&amp;nbsp;Its analysts have business relationships with technology vendors.&lt;/i&gt;&lt;/p&gt;
&lt;/section&gt;</body>
            <description>Don't let batch processing lead to missed opportunities. Build AI systems for continuous data flows that deliver instant decisions, change outcomes and justify the cost.</description>
            <image>https://cdn.ttgtmedia.com/visuals/digdeeper/2.jpg</image>
            <link>https://www.techtarget.com/searchdatamanagement/opinion/Real-time-data-streaming-for-AI-invest-where-it-matters</link>
            <pubDate>Wed, 25 Feb 2026 13:47:00 GMT</pubDate>
            <title>Real-time data streaming for AI: invest where it matters</title>
        </item>
        <title>Search Data Management Resources and Information from TechTarget</title>
        <ttl>60</ttl>
        <webMaster>webmaster@techtarget.com</webMaster>
    </channel>
</rss>
