Welcome

Passionately curious about Data, Databases and Systems Complexity. Data is ubiquitous, the database universe is dichotomous (structured and unstructured), expanding and complex. Find my Database Research at SQLToolkit.co.uk . Microsoft Data Platform MVP

"The important thing is not to stop questioning. Curiosity has its own reason for existing" Einstein



Wednesday, 8 April 2026

Why Data Catalogues Fail (And How Purview Is Quietly Fixing the Industry’s Blind Spots)

Most data catalogues fail for a simple reason: they  assume that documentation alone creates understanding. It doesn’t. A catalogue full of stale metadata, incomplete lineage, and inconsistent tagging is worse than useless and it creates a false sense of confidence. Many organisations have learned this the hard way, investing heavily in catalogues that quickly became digital graveyards.

Purview succeeds where others fail because it treats the catalogue as part of a governance ecosystem, not a standalone tool. Lineage, classification, access policies, and data maps are not optional extras. They are the core of the experience. This integrated approach ensures that metadata is accurate, automated, and actionable.

Another blind spot Purview addresses is operational relevance. Traditional catalogues focus on documentation whereas Purview focuses on control. It doesn’t just describe data as it also governs it. This shift from passive to active metadata is what makes Purview viable at enterprise scale.

Purview also excels in hybrid and multi‑cloud environments, where many catalogues struggle. Its connectors, scanning capabilities, and policy enforcement mechanisms are designed for real‑world estates, not idealised architectures.

Purview is integrated with Fabric which positions it as the governance backbone of the Microsoft ecosystem. As organisations consolidate their data platforms, Purview becomes the source of truth that ties everything together.



Saturday, 4 April 2026

GCRAI and the Rise of GRAICE™: A New Global Framework for Responsible AI Governance

The global conversation around responsible AI has been dominated for years by national strategies, corporate principles, and academic frameworks. But the launch of the Global Council for Responsible AI (GCRAI) and its GRAICE™ framework marks a shift toward something far more ambitious: a unified, cross‑sector, cross‑industry operating system for AI governance. Unlike many initiatives that focus on high‑level ethics, GCRAI positions itself as a mechanism for operationalising responsibility at scale. It’s an attempt to move responsible AI from aspiration to enforceable practice.

What makes GCRAI notable is its global footprint. With representation across dozens of countries and a network of ambassadors, it aims to create a governance ecosystem that transcends borders and industries. This matters because AI risk is not localised. Models trained in one region influence decisions in another. Data flows across jurisdictions. And the consequences of AI misuse rarely stay within organisational boundaries. A global framework is not just desirable, it is necessary.

The GRAICE™ framework, unveiled at Davos, is positioned as “humanity’s operating system for AI.” While the branding is bold, the intent is clear: create a standard that is actionable, measurable, and adaptable. GRAICE™ focuses on transparency, security, accountability, and human‑centric design. But what sets it apart is its emphasis on measurable compliance. Many frameworks articulate principles; GRAICE™ attempts to define behaviours. It seeks to bridge the gap between what organisations say about AI and what they actually do.

Running alongside GCRAI is the G.R.A.C.E. Global Council for AI, which articulates a complementary set of principles centred on human‑centred AI. Their pillars emphasise mission, vision, and the balance between technology, ethics, and humanity. While still evolving, the G.R.A.C.E. principles reinforce the idea that responsible AI is not just a technical discipline but it’s a societal one. They highlight the need for AI systems that enhance human capability rather than diminish it, and for governance that protects people as much as it protects organisations.

Together, GCRAI and G.R.A.C.E. represent a growing recognition that responsible AI cannot be solved by isolated efforts. Organisations need frameworks that are interoperable, globally recognised, and grounded in real‑world practice. They need standards that can be implemented, audited, and adapted as technology evolves. And they need governance models that reflect the complexity of modern AI systems and systems that learn continuously, behave unpredictably, and operate across boundaries.

For data and AI leaders, the emergence of GRAICE™ is a signal. The era of voluntary, principle‑only responsible AI is ending. The next phase is about operationalisation, measurement, and accountability. Whether organisations adopt GRAICE™ directly or use it as a benchmark, its influence will shape how responsible AI is defined, governed, and enforced in the years ahead. This is not just another framework but a part of a global shift toward responsible AI as a shared, enforceable standard.

G.R.A.C.E. is
GROUNDED
RESPONSIBLE
AUTHENTIC
COMPASSION
ETHICAL

Every decision involving AI should align with moral truth, respect for life, and integrity of purpose through moral align




https://www.graceglobalcouncil.com/
https://gcrai.ai/

Wednesday, 1 April 2026

How GRAICE™ and Microsoft’s Responsible AI Standard Shape the Future of AI Governance

As the responsible AI landscape matures, two frameworks are emerging as influential anchors in how organisations think about AI governance. The newly launched GRAICE™ framework from the Global Council for Responsible AI (GCRAI), and Microsoft’s long‑established Responsible AI Standard. Each framework reflects a different lineage, a different worldview, and a different set of priorities. Yet both are converging on the same fundamental truth: responsible AI is no longer a philosophical debate but an operational discipline.

GRAICE™ enters the scene with global ambition. It positions itself as a unifying operating system for responsible AI, designed to be adopted across governments, enterprises, and civil society. Its principles emphasise human‑centricity, societal impact, and global accountability. The tone is intentionally broad because the problems it aims to address are cross‑border data flows, global AI risk, and societal trust that cannot be solved by any single organisation or nation. GRAICE™ is built for the world stage.

Microsoft’s Responsible AI Standard, by contrast, is built for practitioners. It is grounded in engineering realities: data sourcing, model evaluation, transparency requirements, human oversight, and lifecycle monitoring. It is not trying to govern the world; it is trying to govern systems. Its strength lies in its specificity. It tells teams what to do, how to do it, and how to measure whether they have done it well. It is a framework forged in the crucible of product development.

The contrast between the two frameworks is striking. GRAICE™ is expansive, values driven, and globally oriented. Microsoft’s standard is precise, operational, and system‑oriented. One speaks the language of societal responsibility; the other speaks the language of engineering discipline. Yet this contrast is exactly what makes the comparison so valuable. Together, they represent the two halves of responsible AI, the why and the how.

Where the frameworks converge is equally important. Both insist on transparency as a prerequisite for trust. Both emphasise accountability, not as a slogan, but as a requirement for human oversight. Both recognise that AI systems evolve and therefore require continuous monitoring. And both acknowledge that responsible AI is not a one‑off certification but an ongoing commitment. These shared foundations signal a broader alignment across the industry where responsible AI is becoming standardised, measurable, and expected.

The real opportunity lies in how organisations combine the two. GRAICE™ provides the global context , the societal lens, the ethical north star, the cross‑sector alignment. Microsoft’s Responsible AI Standard provides the operational machinery and the processes, controls, and engineering practices that turn principles into behaviour. When used together, they create a governance model that is both globally relevant and locally actionable.

This is where the future of responsible AI is heading. Not toward a single universal framework, but toward an ecosystem of complementary standards that reinforce one another. GRAICE™ sets the direction; Microsoft’s standard provides the path. Organisations that embrace both will be better equipped to build AI systems that are trustworthy, transparent, and aligned with human values, not just in theory, but in practice.




Saturday, 28 March 2026

Series Index Summary: Data Governance, Purview, and Responsible AI

This four‑month series explores the shifting landscape of data governance, Microsoft Purview, and Responsible AI at a moment when organizations are being forced to rethink how they manage, understand, and trust their data. Across the posts, the series traces a clear arc: from the maturing of governance in 2025, through the practical realities of Purview adoption, to the cultural and architectural shifts required to lead in the age of AI.

The December posts set the stage by examining why governance finally became a strategic priority, how Purview’s quieter updates are reshaping the platform, and why AI risks making organizations intellectually complacent without strong data foundations. These pieces frame governance not as bureaucracy, but as the mechanism that makes innovation safe.

I move deeper into strategy and Responsible AI. It explores the predictions shaping 2026, the operational implications of Microsoft’s updated Responsible AI framework, and the evolution of Purview’s classification engine. The AI Is Making Us Dumber series continues here, highlighting the risks of over‑automation and the importance of maintaining human understanding.

I shift into technical depth and organizational reality. It covers SQL Server’s new direction, the strategic value of metadata, and a detailed breakdown of Purview’s February feature updates. The month closes with reflections on why organizations struggle to operationalize policy and how governance must adapt to keep pace with rapidly learning AI systems.

March brings the series to a forward‑looking conclusion. It introduces the concept of contextual governance, examines the architectural convergence of Fabric and Purview, and challenges data leaders to unlearn outdated assumptions. These posts emphasize that leadership in the AI era requires adaptability, transparency, and a willingness to rethink long‑held beliefs.

Together, these posts form a cohesive narrative about where data governance is heading, what Purview is becoming, and how organizations can navigate the accelerating complexity of AI‑driven data estates. I wanted to add clarity in a landscape full of noise and understand that governance is no longer optional, but foundational.

Wednesday, 25 March 2026

Unifying the Data Estate for the next AI Frontier Fabcon Keynote

The Atlanta FabCon keynote was delivered last Wednesday by Amir Netz (CTO and Technical Fellow), Arun Ulag (President, Azure Data), Shireesh Thota (Corporate Vice President, Azure Databases).  It has was recorded. You can watch it here

Session Abstract

As organizations race to deploy generative and agentic AI, the biggest challenge they face is not models, it’s their data estate. Join Microsoft engineering leadership to learn how Microsoft’s databases can be unified through Microsoft Fabric and OneLake, creating a single, governed foundation for analytics, AI, and intelligent agents. Discover why this shift represents a fundamental change in how modern data platforms are built, managed, and scaled for the next AI frontier.

A summary of the announcements.





Sunday, 22 March 2026

What Data Leaders Must Unlearn to Lead in the Age of AI

The hardest part of leading in the AI era isn’t learning new skills, it is unlearning old assumptions. Many of the beliefs that shaped data leadership over the past decade no longer apply. The pace of change, the complexity of modern estates, and the unpredictability of AI systems demand a different mindset. Leaders must be willing to let go of outdated models of control, certainty, and hierarchy.

One of the first assumptions to unlearn is that governance slows innovation. In reality, governance accelerates innovation by reducing risk, increasing clarity, and enabling responsible experimentation. When governance is embedded rather than imposed, it becomes a catalyst rather than a constraint. Leaders who cling to the old narrative will find themselves outpaced by those who embrace governance as a strategic enabler.

Another assumption to unlearn is that documentation equals understanding. In the AI era, understanding comes from lineage, monitoring, and behavioural metadata, not static documents. Leaders must shift from documenting after the fact to embedding governance into the system itself. This requires investment in tooling, automation, and literacy.

Leaders must also unlearn the idea that AI systems can be trusted without oversight. AI is probabilistic, not deterministic. It requires continuous monitoring, not one‑time validation. The organisations that thrive will be those that treat AI as a dynamic system requiring ongoing governance, not a product that can be finished.

Finally, leaders must unlearn the belief that expertise is static. In the AI era, expertise evolves. The best leaders will be those who remain curious, adaptable, and willing to challenge their own assumptions. Unlearning is not a weakness but a leadership skill.



Friday, 20 March 2026

Navigate AI on Your Data & Analytics Journey to Value - Gartner 2026

The Gartner Data & Analytics Summit (March 9–11, 2026, in Orlando) marked a significant shift from AI experimentation to AI industrialization. My post focuses on how governance is no longer a check-the-box activity but the literal engine for AI ROI.

​Here are some collated highlights that interested me.

1. The Core Keynote: Beyond the Hype to ROI

Analysts Adam Ronthal and Georgia O’Callaghan opened the summit by challenging the move fast and break things mentality. They argued that while AI is accelerating, success belongs to those who find a thoughtful approach to speed and direction.

Gartner emphasized that AI adoption follows an S-curve, a slow start, rapid acceleration, then stabilization. We are currently at the steep upward slope. Organizations that don't integrate governance now will face expensive catch-up efforts that turn AI from an asset into a liability.

Gartner categorized firms into three types: AI-First (aggressive), AI-Opportunistic (fast followers), and AI-Cautious (waiting for stability). They noted that regardless of the path, doing nothing is no longer an option.

2. Data Governance: The Move to Adaptive & Autonomous

A major takeaway was that traditional, manual data governance is dead. It cannot keep up with the volume and velocity of AI-driven data.

Gartner introduced the concept of Outcome-Based Governance. Instead of governing all data equally, teams should focus on high-value data products that directly impact AI outcomes.

A new AI-Ready Data Framework focuses on three pillars:

   Alignment: Ensuring data semantics and lineage are clear.

   Qualification: Continuous data quality validation for model training.

   Governance: Enforcing policies during the AI lifecycle.

The Rise of Governance Agents: A top 2026 prediction is that D&A leaders will begin using Data Governance Agents to automate the negotiation and orchestration of data pipelines.

3. AI Governance: Bridging the Trust Gap

The summit highlighted a looming crisis where 60% of organizations are predicted to fail at realizing AI value due to poor integration between data and AI governance.

Gartner warned against Registry-First Governance. Simply listing your AI models in a spreadsheet isn't enough. They called for Continuous Code-to-Cloud Visibility, where governance monitors data as it flows through APIs and AI agents in real-time.

A buzzword at the conference was the Unified Context Layer. To govern AI effectively, you need a layer that connects business meaning to raw data. This allows AI agents to act reliably because they understand the why and how, not just the what.

Gartner predicts spending on AI governance platforms will reach $492M in 2026, doubling to $1B by 2030, as companies realize that compliance is a trust dividend rather than a tax.

4. Responsible AI: Ethics as an Operational Metric

Responsible AI (RAI) moved from a philosophical discussion to a technical requirement.

Gartner warned that critical failures in managing synthetic data (used to train models when real data is scarce) are a major risk to AI governance. Without metadata tracking the lineage of synthetic data, models risk hallucination loops.

The keynote suggested that data organizations are being reshaped into fusion teams where humans and AI agents work together. Responsible AI here means defining clear boundaries of AI involvement in decision-making.

As we move toward Agentic AI (autonomous agents that can take actions), Gartner highlighted the need for explicit transparency capabilities with the ability to audit why an agent made a specific decision in real-time.

In summary by 2027, organizations that emphasize AI literacy for executives will achieve 20% higher financial performance than those that do not. (Gartner, March 2026). In 2026, AI strategy and Data strategy have become inseparable and you cannot scale the former without governing the latter.

Safeguarding the AI Frontier with Microsoft Purview & Fabric Innovations

The speed of AI transformation is accelerating. However, for many organizations, that speed is throttled by a critical concern, that of Data Governance. At the Microsoft Fabric Community Conference this week, Microsoft unveiled a suite of innovations designed to bridge the gap between rapid AI adoption and robust data security. By deepening the integration between Microsoft Purview and Microsoft Fabric, they are providing a secure-by-design foundation for the AI era.
Here is a breakdown of the major announcements.

Data Security: From Protection to Prevention

In an AI-driven world, data oversharing is a primary risk. Microsoft is addressing this by extending Purview’s sophisticated security controls directly into the Fabric ecosystem.

Information Protection Policies: Security admins can now define policies in Purview that automatically enforce access permissions based on sensitivity labels. If a file is labeled Highly Confidential, Fabric respects those boundaries automatically.
 
Data Loss Prevention (DLP) for Fabric is now in preview, DLP policies can identify sensitive information (like SSNs or credit card numbers) as it is uploaded to Fabric. This allows for automatic risk remediation, preventing data leaks before they happen.
 
Trusted Workspace Access allows for secure connections to data sources behind firewalls, ensuring that OneLake remains a secure environment even when pulling from complex network topologies.

Risk Management: Visibility into the Human Element

Governance isn't just about locking down files; it is about understanding user behavior.
 
Purview Insider Risk Management is one of the most significant announcements integrating with Fabric. This allows organizations to detect, investigate, and act on potentially malicious or inadvertent activities, such as mass data downloads or unauthorized sharing directly within the Fabric environment.

Data Discovery & Curation: The Reimagined Governance Experience

Microsoft is moving away from the policing model of governance toward an enabling model. They’ve introduced a reimagined governance experience that is business-friendly and federated.

Unified Catalog & Metadata: Fabric’s built-in metadata and lineage are now seamlessly reflected in Purview. This gives a single pane of glass view across your entire multi-cloud estate, making it easier for users to find the data they need while ensuring it meets quality standards.

Item Tagging is a new feature allowing users to add tags to Fabric items. This significantly enhances discoverability and encourages the reuse of high-quality data assets across the organization.

AI Readiness: Building on a Trusted Foundation

The ultimate goal of these updates is AI Transformation. You cannot have a reliable Copilot if it is grounded in unverified or ungoverned data. By automating discovery, classification, and protection, Microsoft Purview ensures that the data fueling your AI models is:
 Accurate: Through better curation and quality checks.
 Compliant: Adhering to regional and industry regulations.
 Secure: Only accessible to the right people (and the right AI agents).

Final Thoughts

The announcements from FabCon 2026 signal a shift. Data governance is no longer a hurdle to be cleared. It is the engine that allows AI to run safely. For those of us managing data estates, the tighter synergy between Purview and Fabric offers a clear roadmap to innovate with confidence.


Wednesday, 18 March 2026

FabCon and SQLCon 2026

It is the third Fabcon event and first ever SQLCon in Atlanta this week. This week Microsoft didn't just make a small change, they made a larger shift. The convergence of where data lives and what data does has been the holy grail of database management and the biggest hurdle to AI isn't the model itself, it is the data ingest quality and the fragmentation of the estate. The announcements from the conference help answer the current chaos and complexity that exists. This is my take on the key shifts that will matter most when navigating the database landscape.

The Single Pane of Glass Arrives in the Database Hub
For years, we have managed our estates in silos, Azure SQL over here, Cosmos DB over there, and SQL Server on-premises (hopefully via Azure Arc) somewhere else. The new Database Hub in Microsoft Fabric (now in early access) is a game-changer for governance. It provides a unified view to explore and optimize the entire estate. But the real interest is the agent-assisted management. Using intelligent agents to reason over signals and explain why something changed. It keeps the human in the loop but removes the manual drudgery.















Microsoft IQ: The Semantic Layer for AI

One of the biggest announcements is how Fabric is becoming the intelligence layer for the enterprise.
  •  Fabric IQ brings together live business data.
  •  Work IQ pulls in productivity signals.
  •  Foundry IQ captures institutional knowledge.
This is critical because AI agents are only as good as the context they have. By creating a unified semantic meaning, we are finally moving away from hunting for data and toward activating data.

OneLake is Closing the Gap on Silos
The OneLake vision continues to expand with more native mirroring capabilities (SharePoint lists and Dremio are now in preview; Oracle and SAP Datasphere are GA).
The standout for me, however, is the Shortcut transformations. The ability to shape data, like converting Excel to Delta tables, automatically as it connects to OneLake is a massive win for data quality. We know that without good quality data at the start, the AI journey hits a wall. These automated gatekeepers help ensure the lake doesn't become a swamp.

Mission Critical Apps with connected SQL and Fabric
With SQL Server 2025 growing faster than any previous version, the integration with Fabric is no longer a maybe. The announcements focused heavily on a converged platform that unifies transactional and analytical data.
For developers, the new Migration Assistant for SQL databases (using AI to resolve compatibility issues via DACPACs) is a pragmatic approach to modernization.

Beyond the Hype it is easy to get lost in the agentic AI buzzwords. But looking at the technical roadmap from FabCon/SQLCon, the focus is clearly on Usability, Empowerment, and Security.

We are moving toward a world where the database is taking a more active place in the business. Whether we are managing a legacy SQL estate or building a greenfield Fabric environment, the wall between our operational databases and analytics is getting less.

 You can read more about the announcements: 


Sunday, 15 March 2026

Metadata Is Not Optional: The Strategic Value Organisations Still Undervalue

Metadata has always been the unglamorous backbone of data governance, but in 2026 it becomes a strategic asset. AI systems depend on it, automation relies on it, and governance collapses without it. Yet many organisations still treat metadata as an afterthought, something to be documented later, if at all. This mindset is becoming increasingly untenable.

Metadata based on DAMA-DMBOK principles is 'data about data', meaning the descriptive information that defines, structures, and gives context to data so it can be understood, managed, and used effectively.

The rise of AI has exposed the consequences of weak metadata practices. When organisations cannot explain where data came from, how it has changed, or who has access to it, they cannot trust the outputs of their models. Metadata is the connective tissue that links data to meaning, context, and accountability. Without it, even the most sophisticated AI systems become brittle.

Metadata also plays a critical role in operational efficiency. Automated classification, lineage, and policy enforcement all depend on accurate metadata. When metadata is missing or inconsistent, governance becomes manual, slow, and error‑prone. When metadata is rich and reliable, governance becomes scalable.

The organisations that succeed in 2026 will be those that treat metadata as a first‑class citizen. This means investing in automation, stewardship, and tooling that captures metadata at the point of creation and not months later. Metadata is not optional. It is the foundation of trustworthy data and Responsible AI.

Inspirational STEM 1958-1968

From the moment I first understood the meaning of my mum’s , Joan Holt, school motto “Be strong and very courageous” I realised it wasn’t just a phrase she carried; it was a quiet force that shaped her life. Long before women in STEM were recognised or encouraged in the way they are today, she worked in a world of theoretical physics, numerical analysis, and early computing with a determination that still leaves me in awe. Hers was not the loud, celebrated courage of someone who set out to break barriers, but the steady, purposeful courage of someone who simply refused to accept that those barriers applied to her. With it being International Women’s Day last week we celebrated the women who paved the way and that reminds that one of those pioneers was my mum.

Her career reads like a living history of British computing. Her first days were working on IBM mainframes and analysing data, when computers filled whole rooms and printouts were the size of phone books to programming the Ferranti Mk1. She drafted manuals for the Elliott 503 and 4100, and solved problems with nothing but symbolic assembly code. She lived through the evolution of technology as few people did. She worked in rooms where magnetic tapes towered over her, where data meant punched cards and where a single mistake meant repunching a deck. She navigated machines that shook themselves off desks, deciphered the results of calculations that once took over 8 months, and wrote documentation that bridged engineers and the future operators. She was often the only woman in the room, one of only a handful among thousands of men at just nineteen. She simply worked hard proving herself indispensable through intelligence, persistence, and grace.

Today, on Mother’s Day, I think not only of the extraordinary work she did, but of the extraordinary woman she was. A role model who taught me that courage can be quiet, curiosity can be powerful, and that you can shape the world even if you never stand in the spotlight. While the world now celebrates women in STEM more visibly than ever, she lived those values when the path was far tougher and the recognition far thinner. Her achievements may sit in old manuals, early programs, and memories of rooms filled with tapes and valves, but her legacy is alive in me. I am proud beyond words to have been her daughter, and prouder still to share her personal story to help inspire future generations.





Saturday, 14 March 2026

Fabric, Purview, and the New Shape of Enterprise Data Architecture

Fabric has reshaped the Microsoft data landscape by unifying analytics, engineering, and storage into a single experience. But unification alone does not create coherence. The real transformation happens when Fabric is paired with Purview. Together, they form an architecture where data movement, governance, and analytics operate as one system rather than disconnected components.

This convergence matters because modern data estates are too complex to govern manually. Data flows across pipelines, notebooks, semantic models, and AI workloads. Without integrated governance, organisations end up with pockets of visibility rather than a complete picture. Purview provides the lineage, classification, and policy enforcement that Fabric alone cannot deliver.

One of the most powerful aspects of this integration is the alignment between data products and governance. Fabric encourages teams to think in terms of products that are curated, reusable assets with clear ownership. Purview reinforces this by providing the metadata, stewardship, and controls that make data products trustworthy. Governance becomes part of the product lifecycle, not an afterthought.

This new architecture also supports hybrid and multi‑cloud realities. Many organisations are not all using Fabric, nor should they be. Purview’s ability to govern across environments ensures that Fabric becomes a strategic hub rather than a silo. The result is an architecture that is unified, not monolithic but flexible.

As organisations modernise their estates, the combination of Fabric and Purview will become the default pattern. It is not just a technical alignment but it is a governance first architecture for the AI era.

It is FABCON and SQLCON in Atlanta March 16 - 20, 2026. 



Sunday, 8 March 2026

Purview and OneLake Govern tab change

The Purview Hub in Fabric insights have now moved to the OneLake catalog’s Govern tab. The change helps bring governance closer to where the data actually lives, rather than leaving them in a parallel experience that always that wasn't as helpful as it could have been. In the Govern tab, you now see the same posture summaries, recommended actions, and learning resources that were in Purview Hub, but framed within Fabric’s unified governance model. It is a cleaner, more coherent way of surfacing what core information about the health of your data estate.

Functionally, the Govern tab now gives you a consolidated view of governance status, recommended actions, sensitivity and endorsement insights, and links into deeper governance tooling. You can drill into items that need attention, track improvements over time, and understand how your organisation is using Fabric’s governance features. The experience also ties directly into the OneLake catalog, so governance isn’t an afterthought. It is embedded in the same place you explore, classify, and manage data assets.

Microsoft hasn’t yet published a formal retirement date for the Purview Hub. Fabric is now  presenting a single, coherent story about how organisations should understand and manage their data estate.






You can learn more about it here.

Friday, 6 March 2026

Why Strategic Leaders are Pivoting to Contextual Governance

For decades, data governance has been treated as a static discipline with a set of rigid policies laid out in formal frameworks and applied uniformly across the enterprise. But in an era defined by decentralized architectures and the breakneck speed of AI adoption, this one-size-fits-all approach is inefficient and has the potential to increase business risk.

The mismatch between static governance and dynamic data estates is the primary reason why many digital transformation projects stall. It is time to move toward Contextual Governance.

The Governance Friction Paradox

Traditional governance models fail because they are binary. They treat data as a fixed asset rather than a fluid utility. This creates a paradox:

  • Over-governance: Smothering low-risk innovation with unnecessary red tape.

  • Under-governance: Missing the subtle, high-risk nuances of how data is actually used in the wild.

Static rules rely on metadata labels that are often outdated the moment they are applied. Contextual governance, however, shifts the focus from what the data is to how the data is behaving.

What is Contextual Governance?

Contextual governance is a move from policing to orchestration. It is an adaptive framework that evaluates risk in real-time based on the intersection of three pillars:

  1. The Actor: Who is accessing the data, and what is their historical behaviour?

  2. The Environment: Where is the data flowing? Is it a sandboxed R&D environment or a customer-facing LLM?

  3. The Intent: Is the data being used for a routine report, or is it being fed into a model that could leak proprietary logic?

The Strategic Shift: We are moving from asking, Is this data protected? to asking, Is this data protected enough for this specific moment?

Beyond Compliance: The Competitive Edge

For the C-suite and strategic leads, contextual governance isn't just a compliance checkbox. It is a performance multiplier.

  • Agility at Scale: By automating the easy permissions and tightening controls risk is reduced and  the bottlenecks a removed that frustrate engineering and data science teams.

  • AI Readiness: AI systems don't live in a vacuum. A model that is safe in a localized test may become dangerous when exposed to real-world edge cases. Contextual governance provides the guardrails necessary to deploy AI with confidence.

  • Intelligent Foundations: This shift forces a higher standard for metadata and lineage. You are mapping data and mapping the value stream of the entire organization.

The Path Forward

Transitioning to this model requires more than new software; it requires a cultural pivot. We must change how we view governance from firm control to see it as a central intelligent system of the enterprise.

The future of data doesn't belong to those with the thickest rulebooks. It belongs to those who can govern at the speed of business.




Monday, 2 March 2026

Purview Announcements Round‑Up for February Features

February’s Purview updates delivered a series of enhancements that, while not flashy, significantly strengthen the platform’s ability to support enterprise‑scale governance. These updates reflect a clear pattern: Microsoft is investing in the operational realities of governance, not just the conceptual frameworks. The result is a more mature, more capable, and more integrated Purview experience.

One of the standout improvements is the expansion of lineage depth and clarity. Complex estates often struggle with lineage that is either too shallow to be useful or too dense to interpret. The new enhancements strike a better balance, offering more detail without overwhelming users. This is particularly valuable for organisations trying to trace data through Fabric pipelines or hybrid architectures.

Another important update is the introduction of more granular policy controls. Organisations increasingly need to apply nuanced rules across different environments, data types, and sensitivity levels. The new controls make it easier to enforce policies that reflect real‑world complexity rather than idealised models. This is governance that adapts to context, not governance that forces the business to adapt to the tool.

Purview also continues to strengthen its integration with Fabric. As more organisations consolidate their analytics and data engineering workloads into Fabric, the need for seamless governance becomes critical. The February updates improve the consistency of metadata flow between the platforms, reducing gaps and improving trust.

Finally, the enhancements to scanning and classification performance will make a noticeable difference for large estates. Faster scans, more accurate tagging, and improved connector reliability all contribute to a smoother governance experience.

Saturday, 28 February 2026

Microsoft Purview Data Governance – Interactive Experience

This interactive Storylane provides a guided walkthrough of Microsoft Purview Data Governance, showcasing how organizations can establish clarity, accountability, and trust across their data estate using a modern, federated approach. The experience demonstrates how Purview brings together data discovery, governance domains, data products, access workflows, data quality, and estate health into a single, coherent governance platform.

The walkthrough highlights how business and technical users can discover and understand data through the Unified Catalog, using familiar business language, lineage, and context rather than low‑level technical metadata. It shows how data is organized into business domains and data products, helping teams govern data at scale while maintaining clear ownership and accountability.

The Storylane also illustrates Purview’s end‑to‑end governance lifecycle — from requesting access to governed data, through to defining critical data elements, managing data quality rules, and monitoring data estate health. A key theme throughout is Purview’s federated governance model, enabling central oversight while empowering domain teams to own and manage their data within agreed standards and controls.

Overall, the experience positions Microsoft Purview as a system of record for data governance, supporting organizations as they move toward cloud, analytics, and AI by ensuring data is discoverable, trusted, compliant, and ready for reuse.



Watch the video here

https://purviewdatagovernance.storylane.io/share/kfdnjhua9hlh

Friday, 27 February 2026

The Governance Gap and Why Organisations Still Struggle to Operationalise Policy

Most organisations don’t have a policy problem, they have an operationalisation problem. Policies exist, but they’re not enforced, monitored, or embedded into workflows. Governance becomes a theoretical exercise rather than a practical one. Teams know what they should do, but the mechanisms to ensure they actually do it are missing.

This gap often emerges because governance is treated as documentation rather than behaviour. Policies are written in isolation, disconnected from the systems and processes they’re meant to govern. Without automation, policies rely on human discipline, and human discipline is inconsistent at best.

Microsoft Purview helps close this gap by making policy enforcement automatic and auditable. When classification, lineage, and access controls are integrated, policies become part of the system rather than an external expectation. This shifts governance from aspiration to execution.

But technology alone isn’t enough. Organisations need stewardship, accountability, and a culture that treats governance as part of delivery, not a hurdle to clear. Operationalising policy requires alignment across teams, clarity of ownership, and a commitment to continuous improvement.

The governance gap is not inevitable. It’s a symptom of misalignment. When organisations align policy, technology, and behaviour, governance becomes a strategic enabler rather than a compliance burden.



Thursday, 26 February 2026

AI Is Making Us Dumber, Part III: When Models Learn Faster Than Organisations Do

We have reached a point where models can learn faster than organisations can adapt. This creates a dangerous asymmetry: the technology evolves, but the governance, culture, and literacy lag behind. The result is a widening gap between capability and control. Organisations deploy increasingly powerful models without fully understanding their behaviour, limitations, or risks.

This gap is not caused by technology, it is caused by organisational inertia. Many teams still rely on outdated governance processes that cannot keep pace with continuous learning systems. Policies are static, reviews are infrequent, and oversight is reactive. Meanwhile, models evolve with every new dataset, every retraining cycle, and every shift in user behaviour.

The solution is not to slow the models  but it is to accelerate organisational learning. Governance must become continuous, adaptive, and embedded into operational workflows. This means real‑time monitoring, dynamic policies, and stewardship that evolves alongside the data. It also means investing in literacy so that teams understand not just how to use AI, but how to question it.

When organisations learn as fast as their models, AI becomes a strategic advantage. When they don’t, AI becomes a liability. The choice is not technological, it is cultural.

I spoke on the topic at the  Data Toboggan Winter Edition in a session entitled 'Data Literacy: The Human Advantage in an AI World.'



Tuesday, 3 February 2026

SQL Server’s Next Chapter: What the New Release Signals for Enterprise Data Estates

The latest SQL Server release marked a significant shift in Microsoft’s data platform strategy. Rather than positioning SQL Server as a standalone engine, the new version embraces its role within a broader ecosystem, one that includes Fabric, Purview, and Azure AI. This is not just a technical update but a strategic repositioning that acknowledges how modern data estates actually operate. SQL Server is no longer the centre of gravity. It is a critical component in a distributed, interconnected architecture.

One of the most meaningful changes is the deeper integration with governance and observability tooling. SQL Server has always been strong on performance and reliability, but governance was often something organisations had to bolt on themselves. The new release changes that. Enhanced metadata exposure, improved auditing, and richer lineage signals mean SQL Server can now participate more fully in enterprise‑wide governance frameworks.

Hybrid workloads also receive significant attention. Many organisations still run mission‑critical workloads on‑premises while exploring cloud‑native architectures. The new SQL Server release acknowledges this reality by improving consistency across environments. This reduces friction for teams managing mixed estates and makes it easier to apply governance and security policies uniformly.

For data leaders, the message is clear, SQL Server is evolving to support modern architectures rather than compete with them. It’s becoming more transparent, more governable, and more aligned with the needs of organisations navigating the AI era. SQL Server’s next chapter is one built on integration, not isolation.

Image Source: Microsoft

Friday, 30 January 2026

Data Toboggan Winter Edition 2026

It is that time of year again when Data Toboggan is running another 12 hour conference with 3 tracks with speakers from around the world. There are some amazing sessions to learn from. The conference is free to attend as usual. 

I am speaking on something of interest and topical in my lightning talk in The Chalet on Data Literacy: The Human Advantage in an AI World.

AI is accelerating decision‑making across organisations, but it’s also accelerating how quickly mistakes can scale. This session explores how data literacy keeps humans in the loop, prevents over‑reliance on AI, and strengthens judgment, context, and critical thinking. Attendees will see real examples of AI hallucinations, learn how provenance and triangulation protect against bad outputs, and understand why cognitive skills weaken when tasks are automated. They will leave with a practical checklist for questioning AI outputs, a clear view of the risks of low data literacy, and a framework for building teams that use AI responsibly, confidently, and intelligently.



We have our usual Piste Maps with the agenda.






Wednesday, 28 January 2026

World Economic Forum 2026 in Davos Global Council for Responsible AI

At the 56th World Economic Forum 2026 in Davos between 19–23 January 2026 , the Global Council for Responsible AI officially unveiled GRAICE™ (Global Responsible AI Compliance & Ethics). It is designed as humanity’s operating system for AI. Introduced to global leaders and policymakers, GRAICE moves Responsible AI from principle to practice, integrating ethics, governance, compliance, and human-centric design into a unified, scalable framework. 

The framework is an integrated system rather than a collection of policies that are simple and repeatable.

  • Foundational values established non-negotiable ethical and human centred boundaries
  • Seven pillars translate values into operational requirements
  • Assurance tears verify that requirements are met with evidence
  • Governance structures assign accountability and decision authority

 The six foundational grounded values are  

  • Human dignity and autonomy
  • Accountability and governance 
  • Fairness and justice
  • Transparency an explain ability
  • Reliability and security
  • Inclusivity and social benefits

And the seven pillars for responsible AI define what responsibly I must achieve in practise

  • Ethical leadership
  • purpose driven innovation
  • Human centric use
  • responsible implementation
  • AI literacy and workforce readiness
  • Data governance and integrity



Thursday, 22 January 2026

AI Is Making Us Dumber, Part II: When Automation Replaces Understanding

The second wave of AI dependency is more subtle than the first. It is not about hallucinations or bias, it’ is about the erosion of organisational understanding. As AI tools become more capable, teams increasingly rely on them to summarise, interpret, and decide. Over time, this creates a dangerous dynamic: people stop interrogating the underlying data and start accepting outputs at face value.

This shift is particularly risky in environments where data quality is inconsistent or poorly governed. When teams don’t understand the lineage, context, or limitations of the data feeding their models, they lose the ability to challenge results. AI becomes a black box, and decisions become detached from reality.

Governance is the antidote. By enforcing lineage, quality checks, and human‑in‑the‑loop review, organisations ensure that automation enhances rather than replaces understanding. Governance creates the conditions for informed oversight, not blind trust.

The goal is not to reduce AI usage, it is to elevate human capability alongside it. AI should accelerate insight, not diminish expertise. When governance is strong, AI becomes a partner. When governance is weak, AI becomes a crutch.



Monday, 5 January 2026

The Governance Reset: Five Data Strategy Predictions for 2026

Every January brings a wave of predictions, but 2026 feels different. The pace of change in data and AI has outstripped the pace of organisational adaptation, and leaders are beginning to recognise that their existing strategies are no longer fit for purpose. The old model of annual planning cycles, static governance frameworks, and siloed ownership simply cannot keep up with the velocity of modern data estates. This year will force a reset.

Continuous Governance
The first major shift will be toward continuous governance. Organisations can no longer rely on periodic reviews or manual controls. Governance must operate at the speed of data creation, not the speed of committee meetings. Automated lineage, dynamic classification, and policy‑driven access will become baseline expectations rather than advanced capabilities.

Clarity in areas of data Management
Second, we’ll see a rise in data contracts as a mechanism for aligning producers and consumers. Contracts bring clarity to ownership, quality expectations, and change management. They also reduce friction between teams by making responsibilities explicit. This is governance embedded into delivery, not bolted on afterward.

AI‑driven Metadata Enrichment
Third, AI‑driven metadata enrichment will become essential. Manual documentation has never scaled, and 2026 will be the year organisations finally stop pretending it can. Automated tagging, relationship inference, and behavioural metadata will fill the gaps humans never had time to address.

Cross‑functional Stewardship will Mature
Fourth, cross‑functional stewardship will mature. Governance will no longer sit with a single team; it will be distributed across product, engineering, analytics, and compliance. This shift will require cultural change, but it’s the only sustainable model.

Embrace adaptive policies
Finally, organisations will embrace adaptive policies and rules that adjust based on context, sensitivity, and risk. Static rules cannot govern dynamic estates. Adaptive governance will become the new normal.