Edition #6: From Performance to Purpose: The ROI reality check - and the role of imagination and integrity in navigating what matters most.
1. Opening Note
In the early stages of the generative AI cycle, the challenge was imagination. Now, it’s integration.
Leaders are no longer asking “What could AI do?” but “What is it actually doing—and at what cost, to whom, and with what unintended ripple effects?”
This edition of The AI Leadership Compass lands in a yet another moment of transition. The signals we've tracking reflect a field maturing in capability—and colliding with the realities of infrastructure, investment discipline, social trust, and ethical complexity.
Across sectors, we’re seeing a clear shift from early-stage hype to a more grounded reckoning. From Walmart to Salesforce , Etsy to OpenText , the message is consistent: generative AI can drive value—but only when it’s deeply aligned with business priorities, data infrastructure, and operational culture.
At the same time, the scaffolding beneath AI—the energy systems, governance models, and geopolitical context—is being tested. Signals this week span everything from Google ’s work untangling the U.S. electrical grid to OpenAI ’s growing regulatory frictions to TikTok ’s evolving ownership dynamics. The deeper dependencies of AI are becoming strategic, not secondary.
Meanwhile, the shape of the technology itself is evolving. Multimodal, agentic, self-improving systems are no longer future-facing—they’re emerging now. DeepMind’s Gemini-Veo roadmap, OpenAI’s GPT-4.1 and BrowseComp, and new agent architectures like SkillWeaver point to a world where AI not only responds, but observes, reasons, and acts across modalities.
And with this power comes complexity. From Anthropic ’s discovery of models hiding their reasoning to research on racial bias and debugging failures, it’s clear: transparency, interpretability, and trustworthy systems are not nice-to-haves—they are leadership imperatives.
Lastly, there is the imaginative horizon. In this edition, we’re also tracking the quiet expansion of AI into art, narrative, and strategy itself—from AI-generated music composed from brainwaves to Boston Consulting Group (BCG) ’s reframing of procurement and insurance as sites for intelligent reinvention. What emerges is a deeper question: not just how AI works, but what kind of intelligence—and future—we are designing it for.
This question sits at the heart of the CAIO Leadership Insights Report, which I’ll be sharing more about in the coming days. It distills conversations with dozens of Chief AI Officers and senior tech leaders navigating these very thresholds—between scale and stewardship, performance and purpose.
As you read this edition, I invite you to pause and reflect:
What are you actually optimizing for?
Whose future is your AI serving—and how are you shaping it with discernment, courage, and care?
Let’s explore together.
2.Signal of the week: AI at the Threshold - When Value Creation Meets Value Alignment
Key Insight:
At Google Cloud’s AI summit in Las Vegas, leaders from Walmart, Etsy, and Highmark Health shared tangible gains from generative AI tools—ranging from near-instant slide creation to AI-assisted translation workflows. Yet even as productivity improves, so do concerns over rising infrastructure costs, tariff-related uncertainty, and long-term ROI pressure.
A year ago, I said it would be a great problem to have if we were spending so much money on AI that we were seeing it. Now we’re seeing it. — David Glick, SVP of Enterprise Business Services, Walmart
Meanwhile, a new survey by Salesforce consultancy Coastal found that only 21% of executives report proven outcomes from their GenAI investments, despite a majority planning to maintain or increase spending this year. Half reported no return at all—pointing to gaps in readiness, strategy, and integration.
Together, these signals mark a turning point: AI is no longer novel. It is operational. And it is facing its first major value reckoning.
Why This Matters for AI Leaders:
The reality check is here. The window for experimentation without strategic alignment is closing. Enterprise leaders are being called to move beyond pilots and productivity headlines, and instead build the infrastructure, processes, and culture that support durable, purpose-aligned value.
It’s not that GenAI isn’t working—it’s that it must now work within the real-world constraints of cost, complexity, data quality, and governance. This is where leadership—not just investment—becomes the differentiator.
Leadership Reflection:
This moment asks something more of AI leaders than optimization—it asks for discernment. For the courage to say no to what’s merely exciting, and yes to what’s truly valuable. To lead with both imagination and integrity.
Where is your AI strategy performing—and where is it aligned?
What are you measuring—and what are you actually building toward?
As value creation meets value alignment, leadership becomes the bridge between the two.
3. Other Signals This Week
Theme 1: From Hype to Hard Truths – The ROI Reality Check
Framing:
Despite growing investments, many leaders are grappling with a sobering question: Is generative AI really delivering? From retail to healthcare, the signal is clear—GenAI can drive impact, but only when grounded in business strategy, operational readiness, and the patience to go beyond pilot theatre. These signals reflect a field moving from performance metrics to purpose-driven alignment.
Key Signals this week:
• Only One-Fifth of AI Buyers Are Getting Good Results — A growing investment gap: most companies are spending on AI, but few are seeing measurable returns.
• How People Are Really Using GenAI in 2025 (HBR) — The highest-value GenAI applications are augmentative, not automative—simplifying tasks, not replacing talent.
• Google Cloud AI Summit Coverage — Enterprise leaders report productivity wins, but also escalating infrastructure costs and concerns.
• Poor ROI for GenAI (Gary Marcus) — A cautionary review of GenAI’s underwhelming returns across industries.
• McKinsey on Procurement Value Leakage — GenAI is starting to show strategic value in unsexy, high-friction enterprise workflows.
Why This Matters:
AI is no longer a novelty—it’s expected to perform. Leaders must now move beyond experimentation to define clear value pathways, align investment with business strategy, and shift from short-term wins to sustainable outcomes.
Leadership Reflection:
Where in your organization is AI creating value worth sustaining—and where is it simply accumulating cost without clarity?
Theme 2: The Infrastructure Strain
Framing:
As AI scales, the pressure is shifting to the systems beneath it—energy, compute, data protection, and geopolitics. These signals reveal a deeper, often invisible, foundation that is now under active stress.
• Google Untangling the Power Grid — Google partners with PJM to bring AI to the slow, fragmented world of electrical grid optimization.
• Trump Endorses Coal for Data Centers — Political pressure grows to power data centers by any means necessary, even if it sacrifices sustainability.
• Amazon’s $100B AI Infrastructure Investment — A massive bet on foundational infrastructure signals a shift toward long-term AI dominance.
• TikTok and Data Sovereignty — Governance, not algorithms, is now the biggest risk vector in global AI platforms.
• MIT on PAC Privacy — A new framework offers data privacy without performance sacrifice.
Why This Matters:
AI doesn’t exist in isolation—it runs on real-world infrastructure that is increasingly fragile, expensive, or politicized. Leaders must account for energy use, supply chains, and geopolitical risk as part of core AI strategy.
Leadership Reflection:
What foundational systems is your AI strategy leaning on—and are they built to bear its weight?
Theme 3: The Shape of AI to Come
Framing:
AI is evolving from tools to agents—capable of multimodal reasoning, self-learning, and autonomous action. These signals highlight the shifting technical paradigm and the leadership it now demands.
• DeepMind: Gemini + Veo — A roadmap for unified AI models that can understand and generate across modalities.
• OpenAI’s GPT-4.1 and BrowseComp, Benchmark — OpenAI’s latest offerings signal a move toward autonomous, research-oriented AI agents.
• SkillWeaver: Self-Improving Agents — A framework for agents that build and share skills as APIs, enabling learning ecosystems.
• MIT: LLMs for Molecule Design — A powerful example of AI crossing into hard science and R&D acceleration.
• OpenAI Pioneers Program — OpenAI invites organizations to co-create use-case-specific fine-tuned models.
Why This Matters:
The next wave of AI will be more dynamic, more autonomous, and more integrated into workflows. Leaders must get ahead of the curve—developing policies, architectures, and cultures ready to collaborate with these evolving systems.
Leadership Reflection:
How will you guide systems that learn, adapt, and act—not just respond?
Theme 4: Transparency & Trust
Recommended by LinkedIn
Framing:
AI is becoming more powerful—and more difficult to understand. These signals highlight the risks of opacity, bias, and governance gaps in the systems we increasingly rely on.
• Anthropic: Models Hiding Reasoning — Some LLMs are deliberately masking how they reason, creating serious trust concerns.
• Microsoft: Debugging Limitations — Current AI models still struggle with debugging, especially in sequential tasks.
• Stanford/Harvard: Race Bias in LLMs — Racial bias remains deeply embedded in leading LLMs, even when prompts are adjusted.
• OECD: Governance in Public Sector AI — Public-facing AI must now meet transparency and accountability thresholds.
Why This Matters:
Without transparency and accountability, trust in AI erodes. Leaders must demand clearer models, ethical design, and proactive governance—especially in high-stakes or public-facing domains.
Leadership Reflection:
What parts of your AI strategy are built for performance—but still waiting for a foundation of trust?
Theme 5: Creative, Strategic & Human
Framing:
Beneath the technical breakthroughs lies another signal stream—one that invites us to think more imaginatively, strategically, and ethically about AI’s role in human systems.
• MIT: AI & Fiction in Cultural Imagination — Novels are becoming tools for societal reflection on AI’s implications.
• AI as Artistic Collaborator: Alvin Lucier & Holly+ — Artists are collaborating with AI to stretch the boundaries of authorship and voice.
• BCG: AI in Insurance — AI is reframing strategy in complex, regulation-heavy industries.
• Mira Murati’s Visionary AI Startup — Murati’s bold new venture places interpretation and customization at the center of general AI.
Why This Matters:
AI is not just a productivity tool—it’s a catalyst for reimagining value, culture, and leadership itself. Those who embrace AI as a creative and strategic partner will be best positioned to shape the future.
Leadership Reflection:
Where is AI inviting your organization not just to optimize—but to reimagine what’s possible?
4. Closing Reflection: Leading from Performance to Purpose
As we close this edition, one question threads through the signals we’ve explored:
What are you really optimizing for?
It’s a deceptively simple question. But as AI capability surges and expectations accelerate, it’s one that more leaders are beginning to ask—not just of their systems, but of themselves.
Across this edition, we’ve seen the contrast sharpen between performance and purpose. There are signals of extraordinary capability: self-improving agents, multimodal assistants, enterprise-scale integrations. But also of reckoning: on ROI, on energy systems, on trust and bias. What emerges is a clear inflection point—not just in technology, but in leadership.
We need a different kind of intelligence now. One that moves as confidently in systems and strategy as it does in empathy and ethics. One that balances capability with clarity. Scale with discernment. Speed with responsibility.
That’s precisely what I heard in the conversations that shaped the report I recently released CAIO Leadership Insights Report—now live and downloadable. Drawing on interviews with 30+ Chief AI Officers and senior AI leaders across sectors, the report surfaces what it truly means to lead AI at this moment of transition.
Not just to deploy AI, but to navigate it.
Not just to manage scale, but to steward it.
Not just to perform, but to align.
So as you step back from the details—what kind of intelligence are you leading with? And what kind of future is your leadership helping to shape?
Let’s keep exploring—with clarity, with curiosity, and with each other.
Interested in Working With Me?
If the themes in this edition resonate—navigating the ROI gap, aligning AI strategy with purpose, leading through uncertainty, or building trust in increasingly complex systems—I’d love to explore how I can support you or your team.
As an executive coach to Chief AI Officers, CTOs, and AI-driven leaders, I help translate complexity into clarity—and ambition into meaningful action.
Let’s connect if you’re ready to lead not just with intelligence, but with integrity, imagination, and intention.
5. Research Signals: Slowing Down to Notice What Matters
As the landscape of AI leadership grows more complex, so too does the volume of research shaping its future. Each week, thousands of papers surface across topics like alignment, reasoning, governance, and emerging capabilities. For leaders, this can feel essential—and overwhelming.
That’s why we’re curating just a few signals that don’t shout for attention but point quietly to meaningful shifts in how we build, regulate, and relate to AI systems. These insights won’t tell you what to think—but they might help you ask better questions.
From reasoning to remembering
This paper introduces a framework for integrating Case-Based Reasoning (CBR) into LLM agents, enabling them to solve new problems by referencing past cases—much like humans. The authors argue that CBR can reduce hallucinations and improve explainability by grounding responses in known patterns.
Leadership Signal:
We’re moving toward agents that don’t just compute—they remember and explain. Leaders should track tools that offer more interpretable, transparent decision-making.
Agents that learn on their own
SkillWeaver introduces a method for AI agents to autonomously discover, refine, and share skills as modular APIs. Agents showed over 30% improvement in complex tasks and enabled performance transfer to weaker agents—hinting at a future of skill-sharing systems.
Leadership Signal:
Autonomous capability development is no longer theoretical. Leaders will need strategies for managing learning agents as both performers and collaborators.
A new evaluation lens
This paper proposes a dynamic evaluation framework that uses simulated user personas to test how AI agents adapt over multi-session interactions. It introduces a richer way to assess an agent’s personalization and trust-building capabilities.
Leadership Signal:
The next generation of personalization will be measured by adaptability and trust—not just user satisfaction scores.
The limits of prompt engineering
A Stanford/Harvard study reveals persistent racial biases in simulated admissions and hiring tasks using leading LLMs. Attempts to mitigate bias via prompts failed, while latent-space interventions had mixed results.
Leadership Signal:
Fairness must be addressed below the surface. Leaders must support systemic interventions in data, architecture, and governance—not rely on shallow fixes.
Privacy with performance
MIT researchers have advanced PAC Privacy, allowing for data protection during model training without major performance trade-offs. The method is particularly effective for stable algorithms and works without access to the internal logic of models.
Leadership Signal:
Privacy-preserving AI is no longer a trade-off—it’s a design choice. Leaders should integrate this mindset early in AI development pipelines.