Pentaho 11 is here. See what’s new in our most advanced release yet. Read the blog →
In the coming year, leaders will focus on a few key fundamentals to maximize the value and use of their data for AI.
GenAI’s release of ChatGPT in late 2022 kicked off an acceleration like the data market has never seen. Consider this: there have been over 30 significant updates to general purpose LLMs in 2025 alone!
However, this year did see a fundamental and more permanent mindset shift, and that’s related to data management – specifically data foundations. As organizations looked to move AI from experimentation to operations, they ran headlong into core data challenges that have been festering for years. Quality, governance, and accessibility issues, especially with unstructured data, have kept leaders from deploying AI at scale.
This has driven a real and sustainable interest in data fitness, the disciplined ability to discover, structure, govern, and mobilize data across hybrid environments. Enterprises that treat data as a strategic asset (with unified catalogs, lineage, and policy) are unlocking reliable AI outcomes faster; those that didn’t saw stalled pilots, spiraling storage costs, and automation built on sand.
In thinking about the past year, I see a few key trends that leaders need to consider if they’re going to realize their AI goals in 2026.
Shift From “AI First” To “Data Intelligent First” AI does not fix data problems – it magnifies them. This year exposed the gap between AI tooling and the reality of siloed, opaque enterprise data. Hybrid architectures (cloud + on prem), unstructured repositories (email, PDFs, media), and fragmented governance create blind spots that skew insights and decisions. The remedy is not another model; it’s trustworthy data design: consistent schemas, metadata and context, stewardship, and policies that span environments.
Eliminate ROT And Build Data Ecosystems for Agentic Workflows The industry’s “store everything” habit has turned into a budgetary crisis and a strategic risk. Redundant, obsolete, and trivial data (ROT) inflates storage bills while degrading signals for AI systems. Organizations are responding by auditing what they have, classifying what they need, and automating lifecycle management so data stays purposeful. In parallel, as autonomous agents enter workflows, leaders are learning the hard way that agents require structured, traceable inputs with clear lineage and governance – or else speed simply multiplies mistakes.
How Leaders Can Win In 2026 And Beyond Across our work and writing this year, we stressed four commitments through which leaders can carry AI forward with real impact.
Leaders who adopt these practices can move from AI trials to AI impact. They transform hybrid complexity into coherent, governed data flows that reliably power analytics, automation, and intelligent agents. If you’re interested in diving deeper into these topics, please check out these related columns in Forbes. Thanks for reading, and here’s to a more data fit 2026!
Want to get data fit and AI ready? Explore proven ways to simplify data chaos and operationalize governance with our team.
Author
View All Articles
Featured
Simplifying Complex Data Workloads for Core Operations and...
Creating Data Operational Excellence: Combining Services + Technology...
Top Authors
Mauro Damo
Tim Tilson
Sandeep Prakash
Jon Hanson
Richard Tyrrell
Categories
Pentaho Data Optimizer helps Databricks users reduce cloud storage and compute costs by identifying ROT data, automating tiering and remediation, and ensuring the right data stays fast, trusted, and aligned with business value.
Learn More
What is Data Storage Optimization and Why Is it So Valuable Now? Data storage optimization maximizes the value of data by increasing efficiency, cost-effectiveness, and performance of enterprise data storage. Organizations are generating gigabytes of data every hour, while budgets remain fixed or even decreased. This creates budget and management stress for data storage professionals, […]
From record hail and flood losses to rising cyber threats and regulatory scrutiny, DACH insurers are under pressure from every angle. Pentaho helps carriers cut through data silos, automate compliance, and orchestrate real-time workflows so they can protect margins, customers, and trust when storms hit hardest.
In an era defined by climate risk, regulatory scrutiny, and AI accountability, resilience begins with verifiable truth. Pentaho helps insurers build governed “Golden Sources”, unified, auditable datasets with embedded controls, lineage, and explainability, so every claim, policy, and model stands on trusted data.
When ISG calls your platform “Exemplary,” it means something’s working. Pentaho earned top honors for delivering smart simplicity — integrating, governing, and optimizing enterprise data so businesses can run leaner, faster, and more intelligently.