
AI unlocked: How LLMs will make experts out of everyone
José Parra Moyano joins Mike Wade & Amit Joshi to explore how Gen AI empowers business leaders, democratizes analytics, and drives innovation....

by Faisal Hoque, Paul Scade , Pranay Sanklecha Published October 9, 2025 in Artificial Intelligence • 8 min read
Artificial intelligence is reshaping the competitive landscape, and governments are racing to position their economies for leadership. In the United States, the recently announced AI Action Plan signals a decisive shift in policy, favoring deregulation and rapid innovation over precaution and control.
Winning the Race: America’s AI Action Plan marks an important shift in both policy and philosophy. Rather than the government coordinating and safeguarding AI development – as under previous US administrations and as continues to be the case in most other countries – the plan sets out an approach that emphasizes deregulation, private-sector development, and a “try-first” mentality.
In pursuit of this goal, it outlines dozens of federal policy actions distributed across three pillars:
While the plan still includes room for national standards and approaches to evaluating AI tech and usage, such as NIST’s AI Risk Management Framework, the underlying philosophy is striking in its openness. Its goal is to “dismantle regulatory barriers” and support faster and more far-reaching AI innovation than was previously possible.
For multinationals operating in, or seeking to operate in, the United States, these principles create both opportunities and risks.
On the one hand, a minimalist, light-touch regulatory environment will enable businesses to test minimum viable products (MVPs), implement new tools, and bring new products to market more rapidly than ever before. At the same time, with fewer prescriptive federal guardrails, there will be a heightened risk that flawed algorithms or systems that are rushed into production might fail to meet consumer needs or may even act directly counter to their interests. Such outcomes carry significant ethical and brand risks for the company responsible.
As outlined in an I by IMD article ‘From consumers to code: America’s audacious AI export move’, senior leaders cannot afford to ignore the new environment created by the Action Plan, if only because competitors will move quickly to exploit it. Effective engagement requires a systematic approach that works carefully to make the most of the full range of opportunities the plan offers while simultaneously minimizing the risks involved.
Responding effectively to the AI Action Plan requires a dual mindset that encompasses both radical optimism and deep caution.
Responding effectively to the AI Action Plan requires a dual mindset that encompasses both radical optimism and deep caution. In our recent book TRANSCEND , and an accompanying article in Harvard Business Review, we set out two complementary frameworks designed to help companies operationalize this dual mindset while thinking systematically about how to implement AI-driven transformation.
The OPEN framework – Outline, Partner, Experiment, Navigate – helps companies fully harness AI’s enormous potential, while the CARE framework – Catastrophize, Assess, Regulate, Exit – ensures they effectively manage AI’s equally enormous risks.

These frameworks can provide valuable scaffolding for developing potential approaches for implementing AI in any environment. But four components – Partner, Experiment, Catastrophize, and Exit – offer particular value when responding to the US AI Action Plan.
In the OPEN framework, the Partner stage offers a tool for choosing and shaping relationships so you can bridge resource and knowledge gaps – whether what’s missing is , data, or hard-won know-how. The goal here is to narrow the gap between ambition and capability. In the context of America’s AI Action Plan, progress will often run through partnerships. These include conceptual partnerships between government and the private sector – with government coordinating priorities and standards while firms do the building – and practical collaboration among companies to use common open-weight AI stacks across the supply chain, consistent with the plan’s push for interoperability.
Partnerships also make you more resilient. The plan emphasizes the importance of standards, evaluation, and secure supply chains. Working with the groups shaping those norms will help ensure that your compliance evidence is reusable across business units, boards, and regulators. Understanding future export-control expectations can also reduce the chances that critical components or export paths become unavailable.
Aim to make customers feel like you are experimenting alongside them, not on them.
In OPEN, Experiment means running small, real-world trials to answer practical questions, such as “What value does this create?” “What are the risks and costs?” and “What would it take to run it at scale?” The aim is to learn quickly and inexpensively on the way to making an actionable decision about whether to take a program further or kill it. In the regulatory environment created by America’s AI Action Plan, the US market will provide uniquely beneficial conditions for effectively running “AI labs in the wild.” It will be possible to put new features in front of customers more rapidly and with fewer restrictions, dramatically shortening the path from proof-of-concept to working products.

In the CARE framework, Catastrophize means identifying the worst ways an AI system could plausibly fail so that it is possible to prepare for the risk and avoid or mitigate it. With the light-touch, try-first environment envisioned in America’s Action Plan, the responsibility for catastrophizing shifts decisively to businesses. As the government limits its regulatory requirements, businesses need to take up the slack by becoming the primary custodians of responsible AI implementation.
This is not just an ethical obligation. It is also sound business practice. A proactive approach to identifying risks and defining what levels are acceptable and what are not means leaders can approve appropriate plans with greater speed and confidence.
Think of the Exit step as the development of a three-part architecture, with technical, reputational, and legal layers.
In CARE, Exit means determining well in advance of need precisely under what conditions and how you will stop or unwind an AI implementation. While the AI Action Plan includes a range of guardrails and standards, it leaves control over the exit process almost entirely in the hands of businesses. Pre-defined exit plans shorten crises, limit harm, and preserve value, so it is important to treat them as part of the design, not an afterthought.
Think of the Exit step as the development of a three-part architecture, with technical, reputational, and legal layers. The goal is simple: if an AI system misfires, or if public sentiment requires an implementation to come to an end, your teams must know who pulls the plug, what gets rolled back, and how fast normal service resumes.
America’s AI Action Plan represents a watershed moment for global multinationals, offering unprecedented freedom to innovate while demanding equally robust responsibility in implementation. Success in this new landscape requires that companies move beyond traditional risk-reward calculations and embrace a sophisticated dual approach – pursuing transformative opportunities while vigilantly managing risks.
By adopting these complementary twin tracks while focusing on strategic partnerships, rapid experimentation, proactive risk identification, and clear exit strategies, multinationals can position themselves not just to navigate the AI revolution but to help shape its trajectory.

Founder of SHADOKA and NextChapter
Faisal Hoque, founder of SHADOKA, NextChapter, and other companies, is a three-time winner of Deloitte Technology Fast 50 and Fast 500™ awards and a bestselling author of ten books. His latest, TRANSCEND, a Financial Times book of the month and a “must-read” by the Next Big Idea Club, topped USA Today, Los Angeles Times, and Publishers Weekly bestseller lists. His previous book, REINVENT, a #1 Wall Street Journal bestseller, was published in association with IMD. For thirty years, Hoque has driven sustainable innovation, growth and transformation for organizations including MasterCard, American Express, GE, French Social Security Services, U.S. Department of Defense (DoD), U.S. Department of Homeland Security (DHS), PepsiCo, Chase, and IBM. Named among Ziff Davis’ Top 100 Most Influential People in Technology, his work has been featured in Fast Company, Financial Times, MIT Sloan Management Review, Harvard Business Review, Psychology Today, Wall Street Journal, and other leading publications.

Honorary Fellow at the University of Liverpool and a partner at SHADOKA
Paul Scade is an historian of ideas and an innovation and transformation consultant. His academic work focuses on leadership, psychology, and philosophy, and his research has been published by world-leading presses, including Oxford University Press and Cambridge University Press. As a consultant, Paul works with C-suite executives to help them refine and communicate their ideas, advising on strategy, systems design, and storytelling. He is an Honorary Fellow at the University of Liverpool and a partner at SHADOKA.

Founder of The Philosophy Practice and partner at SHADOKA
Pranay Sanklecha is a philosopher, writer, and management consultant focusing on the intersection of technology, ethics, and practical leadership. Formerly an academic philosopher at the University of Graz, Sanklecha’s research on intergenerational justice includes a book published with Cambridge University Press. He now works with businesses to design and implement philosophy-led frameworks that deliver practical value. He is the founder of The Philosophy Practice and a partner at SHADOKA.

September 30, 2025 • by Michael R. Wade, Amit M. Joshi in Digital
José Parra Moyano joins Mike Wade & Amit Joshi to explore how Gen AI empowers business leaders, democratizes analytics, and drives innovation....

July 16, 2025 • by Faisal Hoque in Digital
AI won't contribute to a thriving business unless it's hardwired into your purpose, people, processes, and architecture. Here’s how getting that alignment right can ensure maximum value from AI investments. ...

July 1, 2025 • by Amit M. Joshi, José Parra Moyano, Ram Bala, Natarajan Balasubramanian in Digital
From the untamed rise of semi-autonomous agents to growing concerns around accountability and security risks, AI agents are creating new headaches for CIOs. Here's how tech leaders can navigate the challenges. ...

June 13, 2025 • by Stéphane J. G. Girod, Michael R. Wade in Digital
As AI reimagines work at unprecedented speed, former Sanofi and Nike executive Andrew Kilshaw reveals what makes this revolution different and how leaders can harness its transformative power. ...
Explore first person business intelligence from top minds curated for a global executive audience