The Digital Transformation Playbook
Kieran Gilmurray is a globally recognised authority on Artificial Intelligence, intelligent automation, data analytics, agentic AI, leadership development and digital transformation.
He has authored four influential books and hundreds of articles that have shaped industry perspectives on digital transformation, data analytics, intelligent automation, agentic AI, leadership and artificial intelligence.
𝗪𝗵𝗮𝘁 does Kieran do❓
When Kieran is not chairing international conferences, serving as a fractional CTO or Chief AI Officer, he is delivering AI, leadership, and strategy masterclasses to governments and industry leaders.
His team global businesses drive AI, agentic ai, digital transformation, leadership and innovation programs that deliver tangible business results.
🏆 𝐀𝐰𝐚𝐫𝐝𝐬:
🔹Top 25 Thought Leader Generative AI 2025
🔹Top 25 Thought Leader Companies on Generative AI 2025
🔹Top 50 Global Thought Leaders and Influencers on Agentic AI 2025
🔹Top 100 Thought Leader Agentic AI 2025
🔹Top 100 Thought Leader Legal AI 2025
🔹Team of the Year at the UK IT Industry Awards
🔹Top 50 Global Thought Leaders and Influencers on Generative AI 2024
🔹Top 50 Global Thought Leaders and Influencers on Manufacturing 2024
🔹Best LinkedIn Influencers Artificial Intelligence and Marketing 2024
🔹Seven-time LinkedIn Top Voice.
🔹Top 14 people to follow in data in 2023.
🔹World's Top 200 Business and Technology Innovators.
🔹Top 50 Intelligent Automation Influencers.
🔹Top 50 Brand Ambassadors.
🔹Global Intelligent Automation Award Winner.
🔹Top 20 Data Pros you NEED to follow.
𝗖𝗼𝗻𝘁𝗮𝗰𝘁 Kieran's team to get business results, not excuses.
☎️ https://calendly.com/kierangilmurray/30min
✉️ kieran@gilmurray.co.uk
🌍 www.KieranGilmurray.com
📘 Kieran Gilmurray | LinkedIn
The Digital Transformation Playbook
The 3AM Question Every CEO is Asking About AI
Use Left/Right to seek, Home/End to jump to start or end. Hold shift to jump forward or backward.
AI is no longer a future concern for CEOs. It is an immediate operational question about speed, risk, and execution.
This episode explores why leaders are under pressure to turn AI investment into real, scalable outcomes.
TLDR / At a Glance
• CEO anxiety driven by pace of AI change
• Shift from chatbots to autonomous agents
• Gap between investment and measurable value
• Reliable use in structured, bounded tasks
• Governance and data as scaling constraints
• 90 day plan built on focus and control
The key takeaway is that competitive advantage will come from focused execution, clear ambition, and early governance rather than broad experimentation.
𝗖𝗼𝗻𝘁𝗮𝗰𝘁 my team and I to get business results, not excuses.
☎️ https://calendly.com/kierangilmurray/results-not-excuses
✉️ kieran@gilmurray.co.uk
🌍 www.KieranGilmurray.com
📘 Kieran Gilmurray | LinkedIn
🦉 X / Twitter: https://twitter.com/KieranGilmurray
📽 YouTube: https://www.youtube.com/@KieranGilmurray
📕 Want to learn more about agentic AI then read my new book on Agentic AI and the Future of Work https://tinyurl.com/MyBooksOnAmazonUK
The 3am CEO AI Question
SPEAKER_00CEO Urgent3 AMAI question you must answer. This article explores why so many CEOs are now asking the same question in the early hours of the morning. Are we moving fast enough on AI or are we already falling behind? The issue is no longer abstract interest in new tools. It is about whether leaders are making the right near-term decisions on capability, risk, and organizational readiness before competitive and regulatory effects begin to compound. After reading this article, you will understand what AI agents mean in practical business terms, where value is proving real and where it still breaks down, and how to build a 90-day plan that is defensible to both the board and the business. Why CEOs are losing sleep over AI. The real question is not whether AI matters, it is whether the organization is moving fast enough and making the right decisions early. Surveys of chief executives show that one of the most urgent concerns is keeping pace with technology, including AI. That concern is grounded in reality because AI is now being embedded directly into the systems where work happens rather than sitting outside them as a separate tool. This shift is being driven by rapid platform changes. Large technology providers are expanding multi-step workflows, system connectors, permissions and controls. At the same time, regulation is moving across major regions, which means faster capability is arriving alongside greater uncertainty. The tension is intensified by a gap between investment and outcome. Many leaders are investing aggressively, yet the financial impact remains uneven. AI has therefore become more than a strategic priority. It has become a source of pressure where organizations are spending but are not yet sure they are winning. What an AI agent means in practical business terms. For most leaders, the simplest way to understand an AI agent is to compare it with earlier tools. A chatbot responds to a prompt and produces an output. An agent carries out multi-step work over time, often across systems, files and applications, with some degree of autonomy and the ability to adjust or escalate when needed. That difference may sound subtle, but it fundamentally changes how AI interacts with the business. As soon as systems begin to behave like delegated actors, they stop being simple productivity tools and start becoming part of the operating model. An agent can retrieve context, plan actions, execute parts of a workflow, and interact with systems in ways that affect real outcomes. That increases value potential, but it also expands the risk surface. Errors are no longer contained within a document or single output. They can propagate across systems, decisions, and processes. This is why ownership becomes a governance issue rather than a procurement issue. It is no longer about who bought the tool, it is about who authorized it, what it can access, how it is monitored, and who is accountable when something goes wrong. Where AI is reliable today and where it still fails. AI is now reliably useful in structured and semi-structured tasks where the boundaries are clear. Drafting, summarization, retrieval across connected documents, spreadsheet support, and coding assistance are all improving quickly and can deliver meaningful productivity gains when deployed within controlled environments. These use cases tend to work because the task is bounded, the data is relatively predictable, and human oversight is straightforward to apply. Problems emerge when systems are expected to operate across multiple steps without clear constraints. Failures often stem from weak data quality, unclear business value, poorly defined permissions, and the absence of stopping rules. This pattern appears repeatedly in analyst and enterprise reporting. Many initiatives work well in controlled demonstrations but fail to scale in production. The challenge is not proving that AI can work. It is ensuring that it works consistently, safely, and in line with business objectives. How to choose your level of ambition. Many organizations struggle not because they lack ambition, but because they choose the wrong type of ambition. A practical way to frame this is through three levels survive, compete, and lead. A survive posture focuses on maintaining competitiveness and controlling cost exposure. A compete posture targets specific areas where AI can deliver measurable improvements in performance. A lead posture aims to reshape operating models or customer experience in ways that create sustained advantage. The mistake is assuming that one level is automatically better than another. The right choice depends on what the organization can realistically support. Copying competitors is one of the quickest ways to fail. Large-scale deployments often depend on deep data infrastructure, strong governance, and mature integration capabilities that are not visible externally. The more useful question is not what others are doing, it is what your organization can execute without creating uncontrolled risk or wasted investment. How to find two or three value pools that matter. Most organizations do not need dozens of AI initiatives. They need a small number of use cases where outcomes are material, measurable, and directly tied to business performance. The strongest candidates tend to be repeatable processes with clear outputs and visible friction. Customer support, financial workflows, document production, compliance processes, and internal knowledge retrieval are common examples. The principle matters more than the category. The aim is to identify areas where the value is concrete and the process can be improved without ambiguity. What readiness really means. Readiness is often described as having the right tools or enough data, but in practice it is broader and more operational. It includes data quality, workflow integration, governance structures, and managerial capability all working together. Data must be accessible, consistent and governed so that AI systems can operate on reliable inputs. Workflows need to be mapped in enough detail that leaders understand where AI reads, where it writes, and where decisions must be escalated. Governance must include permissions, logging, monitoring, and clear escalation paths, not just high-level policy statements. Just as importantly, managers need to be able to supervise AI-enabled work. Many organizations focus on training employees to use tools, but underestimate the importance of training managers to interpret outputs, manage risk, and decide when to trust or override the system. Without that layer, even well-designed systems struggle to scale. UK, EU and US differences that change the plan. A global organization cannot rely on a single AI strategy because legal and regulatory expectations differ across regions. In the European Union, the AI Act introduces staged obligations with major requirements expected from August 2026. Recent proposals have created some uncertainty around timelines, which means leaders need to plan for multiple scenarios rather than assume one fixed date. In the United Kingdom, immediate pressure centers on copyright, training data, and licensing, with government publications and parliamentary scrutiny, shaping near-term policy direction. In the United States, the environment remains more fragmented, with federal and state level developments evolving at different speeds. This makes strong internal governance even more important because organizations cannot rely on one external standard to guide every decision. A credible 90-day plan. A credible 90-day plan begins with clarity. Define the level of ambition, select two or three measurable use cases, map workflows in detail, establish a safe data perimeter, build minimum viable governance, run production level pilots, upskill managers, implement a monthly review cadence. The strength of this approach is that it stays small, specific, and governed. It avoids the false comfort of a big transformation program without operational control. Conclusion. The three in the morning AI question reflects a real shift. Capability is accelerating, investment is rising, and outcomes remain uneven. The organizations that move ahead will not be those that experiment the most. They will be the ones that decide clearly, focus narrowly, and build governance early. This concludes the article. You can also read this article on my LinkedIn page where I share regular insights on AI, strategy, and emerging technologies.