Artificial intelligence (AI) isn’t just empowering organizations to deliver more proactive, predictive, and innovative solutions. It’s also exponentially increasing the rate of information being generated today. In turn, this growing tide of information is prompting enterprises to develop more effective data management processes and build more advanced IT infrastructures capable of ingesting and analyzing an increasing number and variety of data sources.
“It’s early days, but even at this stage it’s clear that AI is having an impact on organizational data volumes,” Simon Robinson, Principal Analyst at Omdia, told The Forecast. “Indeed, 87% of IT decision makers now report that AI is driving substantial data growth, according to one of our recent studies.”
According to MIT, the total amount of data generated jumped by a factor of 32X in the decade from 2010 to 2020. That was before AI took off. The impact of AI quickly reached the tipping point of widespread consumer usage, reported Menlo Ventures, and the technology is prompting “habit formation at an unprecedented scale.”
Other researchers report that nearly 4 in 5 organizations and 52% of adults are currently using large language models (LLMs) like ChatGPT. It’s clear that both the volume of data being generated and consumed by people and enterprises is compounding. Adding to this explosion is a new phenomenon: more content appearing on the Web is generated by AI than humans.
In 1776, Benjamin Franklin’s printer churned out dozens of pamphlets. In the late 1800s, tabulating machines were used to count hundreds of census returns. In the 1960s, mainframes stored and accessed thousands of customer records. In the 1990s, relational databases held millions of business transactions. Circa the 2000s, the Web spawned billions of data points, which multiplied with the rise of social networks, video aggregators, internet-connected sensors and the digitization of almost everything. AI is moving the needle an order of magnitude further.
IDC estimated that if the entirety of the world’s data up to a decade ago were imprinted on DVDs, the stack would be large enough to stretch to the moon 23 times or circle Earth 222 times over. Today, it’s estimated that almost 6X as much information is being generated on an annual basis alone, a figure that could balloon by 1000% by 2030.
Each new IT innovation jolted the world’s levels of data generation and content output upward. But experts say that what we’re currently seeing in the age of generative AI is an exponential leap that promises to be several degrees greater.
Chief Innovation Officer at ServiceNow, Dave Wright, aptly wrote for CIO.com: “The data reckoning has arrived.”
Yet, this generational advancement is not just faster, denser, and richer in variety than any boom in information technology seen to date. It fundamentally changes the nature of what enterprises will need to do with information going forward… and will greatly impact how they’ll store, manage and govern it in the years to come.
“What’s notable is that AI workloads are utilizing data across the board,” noted analyst Robinson. “Our research suggests organizations aren’t just using structured information but also all manner of structured and semi-structured data as well.”
Helpful to note is the way in which Generative AI now influences the way that technology leaders look at information oversight and data operations.
For instance, there’s the issue of volume (how much), noting that more data is now being generated than in any prior point in history.
Then there’s the concern of velocity (how fast), keeping in mind that feedback loops from user interactions, iterative processes, adaptive solutions and smart apps generate, consume, and replicate information at speeds that far outstrip older processes.
Teams also face issues relating to variety (different forms of information), noting that structured, semi-structured and unstructured data can now take the form of text, images, videos, audio, 3D data, sensor streams, etc.
And of course enterprises must also consider veracity (accuracy, context, etc.), noting that AI models thrive on clean, labeled, high‑quality data and metadata, provenance solutions, and privacy handling.
“High-quality data management is critical to making AI results more effective,” Alex Almeida, senior product marketing manager at Nutanix, told The Forecast.
“You need tools that help you classify your data and to be able to better see and analyze the information to determine its importance. You also have to ask yourself questions like if this data should be managed and moved to closer sites to better manage LLM processing. Organizations are currently realizing just how important it is to have their data governance and management houses in order.”
Going forward, Almeida said companies need a robust data management strategy and tools. They must understand where data is collected, gathered and housed inside the business. Data sovereignty rules will require continuous data audits. Enterprises need to classify data and organize it so it can easily be fed and moved to a wide variety of applications on-demand.
“If you’re currently dealing with a hodgepodge of data storage solutions, sources, and information silos, it’s important to consider how to unify them,” Almeida advised.
“Any tools or any new data governance paradigms that you can implement to help bring storage and management capabilities into the next generation are going to help with the ROI of AI implementation.”
Robinson sees businesses growing more interested in parallel file system technologies and bringing together a variety of capabilities. “Having GPU and computing resources and underlying storage in close proximity to data could be beneficial.”
Organizations use data for multiple purposes (e.g., reference, process optimization, training). Data lifecycles can frequently stretch far beyond traditional bounds. This creates urgent pressure points on organizations… doubly so the more data that they generate and consume. Some things for IT teams to consider:
Potential Storage Bottlenecks: Not only does storing more varied forms of data (and doing so in larger amounts) tend to come with larger workloads and greater computing demands. It also generally produces rising costs in terms of needs surrounding hardware, storage network bandwidth IO latency and throughput, which can quickly become performance or productivity chokepoints. If data management solutions can’t keep up with rising enterprise demands, it can negatively impact output.
Governance and Data Oversight Concerns: As forms of information available to IT leaders increase in variety and attributes, enterprises must oversee and govern more complex data types. This can lead to concerns surrounding data quality, compliance and privacy, not to mention regulatory oversight and cybersecurity as well. Without good governance practices in place, AI outputs may be unreliable or expose an organization to heightened risk.
More Complex IT Infrastructures: Many enterprises are already paying technology debt in the form of legacy IT tools and systems. In the age of digital transformation and AI automation, firms now face increasing pressure to not only use cloud or hybrid options to enable scale, but also to store certain data on‑premises. That requires companies to deal with concerns relating to interoperability, data movement, and hybrid operating strategies.
Growing Energy and Operations Costs: The more storage, power and computing that enterprises are consuming, the more data center performance and real estate that companies need. Designing an efficient data management infrastructure is now critical, with AI workloads driving data centers to higher densities and creating the need for greater cooling and power demands.
To address rising issues, enterprises are now adopting and building better data management and governance processes and infrastructures across the board.
“Enterprise data tends to be very fragmented and dispersed in nature,” said Robinson.
“Our research suggests that AI really is a hybrid workload that utilizes on- and off-premises data and capabilities… and managing data for AI across this environment can be very challenging across multiple fronts.
“There are issues relating to security, privacy, governance of course, but also ensuring that AI models are updated with new information. This calls for improved integration and understanding between the data layer and the storage layer. There are lots of ways to achieve this, but it’s clear that all of these solutions will require better coordination and cooperation between AI and data engineering teams as well as infrastructure and storage departments.”
Some approaches to consider:
Adopting More All-Encompassing Information Management Strategies – Classifying data earlier on in the collections process and making a point to better tag and categorize or establish ownership of information. Also, better defining data lifecycle policies and maintaining visibility into where data is added, archived and stored. Creating automated pipelines for cleansing, labeling, and standardizing information.
Instituting Better Storage Solutions – Using data platforms and systems that can store different types of information and layer it for quick extraction and analysis as needed. Adopting hybrid storage architectures that combine on-premises practices with public and private cloud solutions to help better optimize for cost, latency and regulatory compliance.
Making IT Infrastructure Upgrades and Optimizations – Designing data centers, networks and cloud operations to better juggle workloads and handle high levels of computing throughput. Using specialized hardware for certain workloads and information locality to reduce latency and data‐transfer costs. Adopting hybrid cloud strategies to enhance privacy, performance and scalability.
Today’s smart, versatile data management tools and systems can be built on top of existing IT solutions. Many IT teams find ways to integrate old and new tools, according to Almedia.
“Much of finding success here comes down to a business’ overall data management practices and how quickly that you can get information from edge to analytics sites,” he said.
“Customers that have implemented solutions like Nutanix Unified Storage tend to find that they're able to adapt much faster because they can more easily move data across the organization, and enable systems to more readily talk to each other. All of a sudden such tools allow for automated workflows that can become real key differentiators.”
Almeida pointed to examples of upgrades, including:
Retrofitting existing on‑premises systems: Adding object‑storage layers and using file systems that support varied types of media. Doing so makes it possible for older relational or file server systems to still serve structured data, while new unstructured data can flow into added layers of IT solutions.
Embracing hybrid and multicloud storage options: Using public cloud storage services and solutions for large unstructured datasets, archives, etc. while using on‑premises or private cloud tools for transactional, sensitive, and high‑performance structured data.
Adopting edge storage to minimize latency: For sensor-powered data or devices that collect data far from central clouds, keeping this information closer to the point of collection before repatriating or summarizing it.
Leverage smart caching, data compression and deduplication techniques: For example, reducing data’s overall footprint by compressing unstructured media, removing redundant copies and using differential storage.
Facilitating smoother data movement and networking: Noting that bottlenecks aren’t always disk-based, but more often surround the movement and transfer of data between apps, networks and systems, enterprises are re‑architecting pipelines, investing in high‑bandwidth, low latency networks, and optimizing data transfer capabilities.
While the nature of data management and governance is rapidly changing, and the amount of information being generated and analyzed is currently skyrocketing, fear not. As noted here, organizations and IT leaders already have access to many of the tools and technologies that they need to more readily adapt.
“When it comes to rethinking data oversight and management policies, there’s still a huge level of concern around privacy and security, and fear of screwing up,” conceded Robinson.
“Organizations also remain concerned about broader cybersecurity threats and a fear around AI displacing humans in the workplace.”
Still, he said that any concerns and costs associated with making the leap to new data management and governance models are dwarfed by the potential drawbacks of being slow to make the adjustment.
“Perhaps the biggest risk is that [companies who are late to the party here] get left behind,” he stated.
“The whole AI space is moving extremely quickly, and the winners will be those that run fast, break things, learn from their mistakes and go again. Our research already shows that data management issues are the biggest challenges associated with enterprise AI, after related expenses.”
Editor’s note: Learn more about Nutanix Unified Storage and Nutanix Enterprise AI technologies.
Scott Steinberg is a business strategist, award-winning professional speaker, trend expert and futurist. He’s the bestselling author of Think Like a Futurist; Make Change Work for You: 10 Ways to Future-Proof Yourself, Fearlessly Innovate, and Succeed Despite Uncertainty; and Fast >> Forward: How to Turbo-Charge Business, Sales, and Career Growth. He’s the president and CEO of BIZDEV: The International Association for Business Development and Strategic Partnerships™. Learn more atwww.FuturistsSpeakers.com and LinkedIn.
© 2025 Nutanix, Inc. All rights reserved. For additional information and important legal disclaimers, please go here.