The Role of the Data Architect in AI Enablement

Pivoting on right-sized foundations, delivered through focused vertical slices, and driven by impact, not just technology.
 •
15 min

https://www.moderndata101.com/blogs/the-role-of-the-data-architect-in-ai-enablement/

Originally published on 

Modern Data 101 Newsletter

, the following is a revised edition.

AI and the Inevitable Hype Cycle

Unless you've been living under a rock or trying to book a GP appointment over the telephone in the UK, it can't have escaped your notice that the hype about AI and its varied uses is reaching fever pitch.

Every PowerPoint presentation worth its salt now has at least three mentions of ChatGPT, two references to "transformational opportunities," and at least one tantalising promise of Agentic nirvana.

Inevitably, following this peak of inflated expectations, we will then enter the trough of despondency before finally emerging, battle-scarred but wiser, into the plateau of productivity. Only at that point might we actually get some work done, rather than simply talking about how AI will fundamentally change every aspect of modern existence while simultaneously rendering most of us unemployed.

Introduction to the Gartner Hype Cycle – BMC Software | Blogs
Evolution of Hype | Source: Gartner

As we navigate this treacherous landscape, trying to sift fact from fiction, it's worth examining the critical but often overlooked role of the Data Architect in laying the groundwork for successful AI initiatives.


The Many Faces of the Data Architect

One reason the Data Architect can be forgotten when talking about AI may be that the term itself is massively overloaded and means different things to different people, rather like how "biscuit" describes a completely different culinary experience depending on which side of the Atlantic you're standing.

A quick survey of job listings reveals a veritable alphabet soup of titles, each purportedly describing a "Data Architect" role but with wildly divergent expectations and responsibilities:

  • There is the Enterprise Data Architect: The wise sage who gazes upon the entire data landscape from their lofty perch, ensuring alignment between business aims and data strategy. The high priest of the data temple, if you will.
  • Then there is the Data Solution Architect: Less concerned with philosophical data musings and more with actually getting things built, these engineers design and implement data solutions often while under the impatient gaze of stakeholders who needed it yesterday.
  • The Data Modeller: The obsessive-compulsive cousin who spends hours ensuring every relationship is correctly defined as aggregation, composition, or association. Thrives in environments with clear rules and boundaries.
  • The AI Architect: The bridge between the traditional data world and the shiny new realm of data science and AI. Speaks both languages but is viewed with mild suspicion by purists from both camps.
  • The Cloud Data Architect: The moderniser who has never met an on-premise solution they couldn't migrate to the cloud, preferably with minimal downtime and maximal handwringing from the security team.
  • The Data Product Architect: The relatively recent addition to the Data Architect pantheon who views life as an interconnected mesh of finely workshopped products, each perfectly aligned to solving a single use-case. If it isn’t distributed or federated, they aren’t interested.

The common thread running through all these roles is a responsibility for designing and maintaining the scaffolding upon which effective data systems and, by extension, AI systems are built. From defining data strategy to implementing governance frameworks, from creating data models to facilitating knowledge graph development, these professionals are the unsung heroes working behind the scenes to ensure that Agentic coffee maker doesn’t mix your matcha latte up with your mocha frappe.


The DIKW Pyramid and the AI Hierarchy of Needs: A Framework for Understanding the Data Architect's Role

Before we discuss how Data Architects can enable AI, it's worth establishing a conceptual framework that helps position their role and the broader considerations we need to keep in mind for AI to be successful.

Two complementary models prove particularly useful here: the classic DIKW (Data-Information-Knowledge-Wisdom) Pyramid and the more recent AI Hierarchy of Needs.

The DIKW Pyramid has been a staple of information science since the 1980s, describing how:

  • Data consists of raw, unprocessed facts and figures.
  • Information emerges when data is organised and given context.
  • Knowledge appears when information is interpreted, synthesised, and applied.
  • Wisdom arises from the holistic application of knowledge, experience and judgement.
Data Connectedness
Interconnectedness: Bridge to Evolution from Data to Wisdom | Source

📝 Related Reads
Back to Basics of Knowledge Management

Meanwhile, the AI Hierarchy of Needs (a clever riff on Maslow's psychological hierarchy and, in this version, inspired by that used by Shopify) outlines the foundational requirements for successful AI implementations.

While traditional AI Hierarchies of Need tend to emphasise tools and technologies, placing AI and deep learning at the pinnacle, Shopify's Data Science Hierarchy of Needs takes a more impact-focused approach that aligns better with the practical role of the Data Architect:

  • Collect and Model: Creating the foundation through data acquisition, platform development, pipeline builds, data modelling, and cleansing.
  • Describe: Using the data to gain a baseline understanding through reporting, dashboards, metrics, segmentation, and exploratory analysis.
  • Predict/Infer: Applying more advanced techniques like statistics, causal inference, and machine learning to address more profound questions.
  • Prescribe: Using insights to recommend specific actions based on analysis and experimentation.
  • Influence: Making an actual impact on the business through whatever means necessary, regardless of the technical sophistication
What's striking about this approach is that it places impact, not technology, at the pinnacle.

Data Architects operate primarily at the lower levels of this hierarchy, establishing the foundation that makes everything else possible. While data scientists and ML engineers might focus on prediction and prescription, data architects ensure that reliable, high-quality data is available first.

This is why the Data Architect's role is so critical. In the rush to climb to the top of these hierarchies, to create influence and prescribe solutions, organisations often underinvest in the foundational layers of collection and modelling. It's rather like trying to build a palace without first ensuring solid foundations and structural integrity. The resulting palace might look impressive in the architectural renderings, but it will inevitably collapse when constructed.

As Shopify's approach reminds us, the most advanced techniques aren't always necessary to create impact. Sometimes, a simple, well-structured dataset with clear documentation can be more influential than the most sophisticated neural network built on shaky data foundations.


The Opera Cake Approach: Vertical Slicing for AI Success

However, while strong foundations are essential, there's a significant risk in the other direction too: spending so much time perfecting the foundational layers that you never actually deliver value. This is where we can learn from an elegant French pastry and a brilliant software development analogy.

About ten years ago, I lived in Blackheath village in London, where I regularly visited a French bakery called Boulangerie Jade. My favourite patisserie there was the delight of their opera cake: a meticulously crafted confection of layered almond sponge, coffee buttercream, and chocolate ganache, each layer distinct yet harmoniously integrated with the others.

In 2003, Bill Wake used this type of layered cake as an analogy for software development. He wrote:

"Think of a whole system as a multi-layer cake, for example, a network layer, a persistence layer, a logic layer, and a presentation layer. When we split a story, we're serving up only part of that cake. We want to give the customer the essence of the whole cake, and the best way is to slice the cake vertically through the layers. Developers often have an inclination to work on only one layer at a time (and get it 'right'), but a full database layer (for example) has little value to the customer if there is no presentation layer."

This analogy applies perfectly to AI enablement. Data Architects must resist the temptation to build comprehensive, enterprise-wide data models or governance solutions before allowing any AI initiatives to go ahead.

Instead, they should collaborate with business stakeholders and data science teams to deliver "vertical slices" that include everything necessary for specific, high-value use cases.
📝 Related Reads
A.
Vertical slicing embodied by data products: Directly coupled with Data/AI use cases | Impact on Lineage
B.
Vertical Infrastructure Slices of Data for Specific Use Cases | How Data Becomes Product

Data Products are vertical slices of the data architecture, cutting across the stack, from data sources and infra resources to the final consumption points (for representation) | Source: How Data Becomes Product

Each vertical slice should include just enough of each layer:

  • Sufficient data governance for this domain
  • The right level of data quality for this specific model
  • Appropriate security controls for this particular application
  • Just enough clarity around the domain model and shared vocabulary, and definitions
  • and so on.

As more slices are delivered, patterns will emerge that can be standardised and expanded into more comprehensive capabilities.

How Data Becomes Product
How patterns emerge through vertical slices: A view into sequential development of patterns across use cases in source, intermediate, and consumption (Data/AI Apps) layers | Source: How Data Becomes Product

With this balanced approach in mind, let's explore ten practical ways the Data Architect can enable successful AI implementations, moving methodically through these hierarchical levels from foundation to execution, but always with an eye toward delivering complete vertical slices.


10 Ways a Data Architect Can Enable AI

1. Baseline Assessment and AI Readiness Evaluation

Before diving headfirst into the latest AI trend, Data Architects should ensure the organisation understands its current state. This means conducting a brutally honest assessment of data maturity and AI readiness.

Rather than boiling the ocean, start by identifying specific business units or functions that already have relatively mature data practices. Map their existing datasets, assess quality, and determine whether they have the foundational elements required for AI: sufficient volumes of relevant data, basic governance, and clear use cases.

Simultaneously, engage with business stakeholders to identify immediate challenges that could benefit from AI solutions. The goal isn't to produce a 300-page report that will gather digital dust in some forgotten SharePoint folder, but rather to quickly identify low-hanging fruit and problems where AI could provide tangible value with existing data assets.

2. Business Context and Process Transformation Ideation

Once a potential use case has been identified, the Data Architect must become temporarily obsessed with understanding the business domain and processes involved. This task is sadly neglected in many AI initiatives that rush headlong into technical solutions. This process should be conducted with business and domain experts, including any Business Architects the organisation may have as part of a wider EA capability.

It involves:

  • Domain Knowledge Acquisition: Immersing oneself in the specific business domain, whether it's mortgage underwriting, supply chain optimisation, or customer service operations. This means abandoning the comfortable confines of technical jargon and learning to speak the language of the business users, however painful that may be. Document the semantics in a domain-level glossary and dictionary, create conceptual and logical models using those semantics that cover just enough to inform the use-case solution design later.
  • Business Process Archaeology: Excavating the current business processes, often discovering ancient workflows that have evolved through years of ad-hoc adaptations and workarounds. Document the as-is state without judgment (yet).
  • Target State Workshop Facilitation: Conducting design thinking sessions that bring together subject matter experts, end users, and technologists to envision transformed processes. What could a mortgage approval process look like when reduced from 30 days to 3 minutes through AI-powered document analysis? What's the right balance between automation and human oversight?
  • Solution Co-Creation: Adopting a product management mindset to design solutions that balance technological possibilities with practical realities. This includes creating mock-ups, journey maps, and prototypes to test assumptions before writing a single line of code.

The output should be a clear vision of how the business process will be transformed, with specific touchpoints where AI will add value and the data and information flows necessary to support it. It should not be a nebulous promise to "use AI to make things better" that leaves everyone scratching their heads about what that actually means.

📝 Related Reads
How to Build Data Products - Design: Part 1/4 | Issue #23

3. AI Solution Appropriateness Analysis

Despite what the army of vendors knocking at your door might suggest, not every problem requires an LLM. As a result, we need to determine which analytical approaches are most appropriate for the problem at hand.

Critically, this analysis must be conducted in close collaboration with the Data Science and AI teams and may overlap somewhat with the Solution Co-Creation phase of the previous step. The Data Architect brings expertise in data structures, quality requirements, and enterprise integration; the Data Scientists bring expertise in algorithms, model characteristics, and analytical approaches. Neither group alone can make the optimal decision.

Together, create a simple decision framework that helps stakeholders understand when to use:

  • Traditional analytics and BI (still perfectly adequate for many problems)
  • Classical machine learning models (for structured data prediction)
  • Computer vision systems (for image and video analysis)
  • Natural language processing (for text and speech)
  • Generative AI and LLMs (for content generation and complex reasoning)

Remember that the simplest solution is often the best. If a problem can be solved with a well-crafted SQL query, there's no need to deploy a transformer-based neural network that requires a small power station to run.

Shopify Data Science hierarchy of needs pyramid showing from top to bottom in increasing size:  Influence, prescribe, predict/infer, describe, collect and model
Shopify’s Data Science Hierarchy of Needs | Source: Shopify Engineering

As Shopify's hierarchy reminds us, impact matters more than technological sophistication.

4. Data Governance, Context Mapping, and Compliance Architecture

Different AI systems have different governance requirements. Traditional ML models reliant on structured data are more sensitive to data quality issues, while LLMs require governance around knowledge management and business contextual metadata.

Data Architects must ensure that governance frameworks extend to these AI-specific concerns while also addressing regulatory requirements. However, consistent with our opera cake approach, focus on creating just enough governance for each vertical slice rather than attempting to build a comprehensive enterprise data governance framework from the outset.

Map data needs to governance implications by considering:

  • For ML models using structured data:
    • How will data quality be measured and maintained?
    • What cleaning or preprocessing is required?
  • For LLMs and generative AI:
    • How will you provide business context through metadata?
    • What knowledge management systems will ensure the model has access to accurate, up-to-date domain knowledge?
    • Are there information architecture requirements such as ontology or knowledge graph development to provide semantic understanding?
  • What are the data privacy implications for model training and deployment?
  • What level of transparency and explainability is required for this particular use case?
  • Are there audit and lineage requirements for high-risk applications?
  • What bias detection and mitigation strategies should be implemented?

Develop just enough governance to ensure appropriate context, understanding, and compliance without stifling innovation. For each AI initiative, create a minimal viable governance framework that can be expanded as the solution matures, rather than insisting on full-scale governance from day one.

📝 Related Reads
How to Create a Governance Strategy That Fits Decentralisation Like a Glove

5. Data Technical Architecture

AI models require data, often lots of it, and ensuring the correct data is available at the right time is a critical architectural concern.

Assess:

  • What data sources are needed, and how will they be integrated? Do we need to move data, or can we provide access via APIs or data virtualisation?
  • If you are working in a Data Mesh/Data Product environment, are there existing Data Products we can source information from or do we need to build new (aggregate) ones? Can we deliver the output for this use-case as another product that will support multiple use-cases?
  • How quickly does source data change, and how sensitive is the model to these changes?
  • What data latency is acceptable for different types of decisions?
  • What level of data quality is required, and how will it be maintained?
  • How will historical data be managed for training purposes?

Avoid over-engineering data pipelines to provide real-time data when daily or weekly updates would suffice. Conversely, ensure that time-sensitive applications have the necessary infrastructure for low-latency data access. The goal is to create a technical architecture that is fit for purpose rather than unnecessarily complex or expensive.

6. AI Technical Architecture

The technical infrastructure for model training and deployment is often underestimated until it becomes a critical bottleneck. Data Architects should proactively work with AI and ML Engineers to support and design a sustainable AI Technical Architecture that supports both initial model development and ongoing operations.

It should address questions such as:

  • How frequently will models need retraining?
  • What data volumes are involved, and how will training data be stored and accessed?
  • How will feature engineering be standardised and version-controlled?
  • What computational resources are required, and are they available internally or via cloud providers?
  • Will inference happen offline (batch processing) or online (real-time)?
  • What are the latency requirements for model responses?
  • How will the system scale during peak usage periods?
  • What fallback mechanisms exist if the model fails or degrades?

Create reusable patterns that incorporate best practices for model versioning, experiment tracking, and deployment automation. By establishing these patterns early, you create a foundation for efficient scaling as AI adoption grows across the organisation.

7. Security and Access Control Design

AI systems often process sensitive data and can potentially expose information in unexpected ways. This is a critical cross-cutting concern across all aspects of solution design. In partnership with CISOs or equivalent teams, Data Architects must ensure appropriate security controls at all levels.

Consider:

  • Role-based access controls for model training, deployment, and usage
  • Encryption for data at rest and in transit
  • Privacy-preserving techniques such as differential privacy or federated learning, where appropriate
  • Monitoring for potential data leakage through model outputs
  • Authentication and authorisation frameworks for API access to models

Design security as an integral part of the architecture, not as a bolt-on afterthought when the CISO starts asking uncomfortable questions the week before launch. Security by design ensures that AI systems can be trusted not just for their accuracy but also for their handling of sensitive information.

8. Model Monitoring and Drift Detection

Models degrade over time as the world changes around them, a phenomenon known as model drift. This monitoring function ensures ongoing quality and maintains the credibility needed for the solution to be a success. Data Architects should work with their AI counterparts to design and support monitoring systems that detect and alert on various types of drift.

Things to consider include:

  • Data drift (changes in the distribution of input data)
  • Concept drift (changes in the relationship between inputs and outputs)
  • Performance drift (degradation in model accuracy or other metrics)
  • Operational issues (latency spikes, resource utilisation, etc.)

Design architectures that not only detect these issues but also facilitate remediation, such as automated retraining triggers or fallback to simpler models when performance degrades. A well-monitored system can maintain its value over time, even as the world around it changes.

As before, create reusable patterns that incorporate best practices and can be leveraged by other projects later on.

9. Consumption and Feedback Loop Design

AI doesn't exist in isolation; it must integrate with existing systems and workflows, where insights are translated into actions. The Data Architect should ensure the solution design is fit not just for model development but also for consuming AI outputs and the all-important feedback loop.

Consider:

  • Who or what will consume the model output? Humans through dashboards or reports? Other systems via APIs? Autonomous agents via MCP servers?
  • Can decisions be automated based on model outputs, or do we need a human in the loop to make or verify them?
  • How will users or systems provide feedback on model accuracy and performance?
  • What metrics will determine model success, and how will they be tracked?
  • How will the system capture and incorporate this feedback to improve future iterations?

Design the entire feedback ecosystem, both for business impact and model improvement. This might include developing lightweight annotation tools for users to correct model outputs or implementing A/B testing frameworks to compare model versions. Remember that effective feedback loops transform a one-time analytical exercise into a sustainable system that delivers ongoing value.

10. Outcome Measurement and Value Tracking

AI initiatives should ultimately deliver business value, not just technical achievements. Data Architects must design systems that track and communicate this value effectively.

Establish:

  • Clear business metrics tied to each AI initiative and how these metrics, in turn, affect broader corporate metrics.
  • Mechanisms to measure improvements attributable to AI.
  • Dashboards or reports that communicate value in business terms, not technical jargon.
  • Processes for reviewing outcomes and adjusting strategies accordingly.

Bake measurement into the architecture from day one, rather than scrambling to quantify value after the fact. This might include A/B testing frameworks, value attribution models, or integration with existing business performance tracking systems. By explicitly connecting AI initiatives to business outcomes, you create a virtuous cycle where success breeds further investment and adoption.

📝 Related Reads
*(for 8, 9, 10) How to Build Data Products - Evolve: Part 4/4 | Issue #30

Final Note: The Pragmatic Path Forward

The journey to AI enablement isn't a single grand transformation but rather a series of practical, incremental steps guided by a clear vision and grounded in business reality. The DIKW Pyramid and the AI Hierarchy of Needs provide valuable frameworks for understanding this journey, highlighting the critical foundational work that must be done before more sophisticated AI capabilities can be realised.

The Data Architect plays a crucial role in this journey, not by promising AI-powered unicorns that deliver infinite ROI, but by methodically building the foundations that make meaningful AI applications possible and working as part of a team to provide complete vertical slices of our Opera cake. In this way, we can ensure that AI initiatives deliver value quickly while gradually building toward more comprehensive capabilities.

The most successful Data Architects in this space will be those who can bridge the gap between hype and reality, between technical possibilities and business needs. They'll focus not on boiling the ocean but on identifying specific problems where AI can deliver tangible value, then systematically addressing the architectural requirements to make those solutions sustainable.

As we navigate the AI hype cycle, Data Architects who take this grounded approach will deliver real value while others wax lyrical about the theoretical possibilities. And when we finally reach that plateau of productivity, they'll be the ones with working systems rather than PowerPoint slides.

After all, in the wise words that no tech consultant has ever actually uttered,

"It's better to have a simple solution that works than a brilliant one that doesn't."

Thanks for reading Modern Data 101! Subscribe for free to receive new posts and support our work.


MD101 Support ☎️

If you have any queries about the piece, feel free to connect with the author(s). Or feel free to connect with the MD101 team directly at community@moderndata101.com 🧡

Author Connect

Author Bio Card
Find me on LinkedIn 🤝

From The MD101 Team 🧡

State of Data Products: First Quarter, 25’

The quarterly edition of The State of Data Products captures the evolving state of data products in the industry. With the Q1: 2025 Report, we’re capturing the whirlwind of AI-natives that have taken the data product space by storm! If you’ve not already tapped into these insights, get them now!

Access Q1 25' Edition

Continue reading

The Fundamentals of Infrastructure as Code in Data Engineering
Data Platform
16 mins

The Fundamentals of Infrastructure as Code in Data Engineering

Universal Truths of How Data Responsibilities Work Across Organisations
Data Strategy
19 min

Universal Truths of How Data Responsibilities Work Across Organisations

Data Quality: A Cultural Device in the Age of AI-Driven Adoption
Data Strategy
10 min

Data Quality: A Cultural Device in the Age of AI-Driven Adoption