

I’ve been recording the Building with AI: Promises and Heartbreaks podcast for a while now, talking with people who are actually building things with AI, not just posting hot takes about it.
My recent conversation with Glen McCracken was one of those “I wish more boards could listen to this” episodes. (episode available here
Glen is the Chief Product and Technology Officer at Lantum, a health-tech scale-up helping the NHS and other providers manage their workforce. He’s been working with what we now call “AI” for nearly 30 years. At one point he was literally working with the creators of R. He’s also one of my favorite contrarian voices on LinkedIn, and the only person I know who can explain AI adoption with a cartoon and a single sentence.
In the episode, Glen described AI as:
“overhyped and underestimated.”
That framing stuck with me. So in this post, I want to unpack a few of the big ideas from our conversation and translate them into concrete actions for data, analytics and AI leaders.
Glen spends a lot of time calling out the noise machine around AI: vendors exaggerating, analysts inventing new categories, consultants telling you that you are already behind and must “move faster”.
He likes hype up to a point. Hype gets boards asking questions. It gets budgets allocated. It creates curiosity.
The problem starts when the question is:
“How do we shove AI into our organization?”
instead of:
“Where are we in pain, and could AI help us do this better?”
AI on LinkedIn now feels like Instagram for enterprise technology
On Instagram, everyone is in Bali, eating perfect food, with perfect kids, in perfect light. Meanwhile your real life is: breakfast with your kids, Slack, six Zoom calls, and maybe some cold coffee.
On LinkedIn, every AI demo looks magical. Every company claims “we transformed X in 30 days with agents”. If you’re a leader sitting in the middle of a very real, very messy organization, you start to think:
“Is my AI journey wrong? Is my company behind? Am I missing something?”
Glen’s answer is basically: you are seeing the highlights reel, not the truth. The truth is much closer to what we wrote about in “Sorry for the mess” – everyone’s data is messy, and it’s Okay and Stop saying “Garbage In, Garbage Out”, no one cares
You don’t see:
The goal is not to opt out of AI because of the hype, and it’s not to chase the hype either. It is to use the hype as air cover while you work on more boring, durable things: foundations, people, data, and actual business problems.
One of Glen’s most useful lenses is how he classifies projects:
“There are push projects and pull projects.”
Push projects, in Glen’s experience, “are really hard and invariably will fail”.
Pull projects are where the magic happens.
If you listened to our earlier episode with Meenal Iyer and read the write-up (Building an AI-powered Intelligent Enterprise), you’ll recognize the pattern: the best ideas come from the front line, not from the ivory tower.
Glen added another useful concept: moving your organization from “subconsciously incompetent” to “consciously incompetent” about AI.
Right now, in many companies:
So his advice is simple:
This is also where managing upwards comes in.
If your board is pushing you with “we need to do more with AI”, you often need to:
It is more honest to say:
“Yes, we’re on it. Here’s how we’re surfacing bottom-up opportunities and validating them.”
than to spin up three impressive but doomed pilots that will be in the “95% that fail”.
If you want a broader framework for this, The right way to set your data & analytics strategy goes into how to anchor D&A (and now AI) in real business priorities instead of tool shopping.
We talked about one of the hottest fear zones: SDRs and sales automation
In one of his examples, Glen mentioned the classic pattern: you bring in AI to automate sourcing and outbound sequencing, and SDRs immediately ask:
“Is this thing going to replace me?”
Glen’s answer comes from years in robotic process automation (RPA). In that world, there is a line he referenced:
“Automation is all about taking the robot out of the human, so it’s taking those highly repetitive activities out of the human.”
You don’t eliminate the role. You change its shape:
We see something similar at Solid. We use tools like Amplemarket for outbound. Instead of writing 200 semi-personalized emails, a human defines:
Then the system generates and executes the outreach, with actual tailoring at scale.
The result:
None of this makes the fear go away automatically. That fear is human. Which means you have to design for it:
When people feel something is being done to them, they resist. When they feel they are co-designing it, they lean in.
My favorite part of the conversation was a concrete story from Glen’s previous role leading data, analytics and AI at a large fintech.
The sales team had a familiar problem:
Salespeople were very clear on their pain:
“There has to be a better way. We want to spend more time selling and less time doing admin.”
Glen’s team built what was essentially a RAG-powered “client 360” assistant
A few things stood out:
There was also one unexpected downside
Senior leaders realized that, previously, they could infer a salesperson’s thinking and skill from how they wrote escalations. Once AI standardized everything, that signal disappeared.
Their solution was clever: since they already had transcripts and emails flowing into the system, they built a lightweight coaching tool. It gave feedback to both the salesperson and their manager on things like:
Same data, different product, different purpose.
There is a pattern here that echoes a lot of what we’ve written before, especially in Beyond Efficiency: How AI is Redefining Data Analytics
And, crucially, Glen kept repeating: the salesperson is still responsible.
You do not want to hear: “Well, the AI wrote it, so if it’s wrong, that’s not on me.”
If you work in data, you’ve probably heard (or said):
“We can’t do AI until our data is clean.”
Glen has lived through the opposite.
In that same fintech, they were stitching together systems from more than 40 acquisitions. Different CRMs, billing tools, ticketing, you name it. The temptation was:
“Let’s fix the data first. Then we’ll build the cool stuff.”
Instead, they:
The result: they effectively crowdsourced data QA
The people who knew the data best – the ones using it every day – now had a reason and a channel to say:
“This says they’ve been a client for three years. It’s at least seven.”
That feedback then improved the underlying pipelines for everyone.
Glen’s main regret: they should have started earlier
This lines up almost perfectly with what we argued in “Sorry for the mess” – everyone’s data is messy, and it’s Okay and Stop saying “Garbage In, Garbage Out”, no one cares
That requires humility from the data team:
Later in the conversation, I asked Glen to bust an AI myth in one sentence.
He went straight for this one:
“AI is really easy to implement. Like it really is… the myth I would bust would be that AI is hard. It’s not hard.”
If a vendor tells you “AI is hard”, his suggestion is simple:
Find one who genuinely thinks it is easy – because they’ve actually done it before.
His view on org design matches that.
He likes Charles Handy’s old idea of federated organizations and applies it to AI:
You didn’t appoint a Chief Email Officer when email showed up. You taught everyone how to use it.
That doesn’t mean there is never a case for a central AI team. Sometimes you really do need tight governance, shared infrastructure and consistent standards. At Solid, our platform team exists for a reason.
But if all AI energy is concentrated in a single “tower”, you get:
If, instead, AI is part of how every CxO does their job, you get:
If you are a CDAO, CPO, CTO or Head of Data reading this while your LinkedIn feed screams “You’re already behind on AI”, here is how I’d summarize Glen’s playbook.
1. Start from pain, not from models.
Build your roadmap from pull: real, repeated pain that business teams are begging you to fix. “We spend 10 hours a week on X” is a better starting point than “We should have an agent strategy.”
2. Use hype as cover, not as a spec.
The noise machine buys you air cover to invest in literacy, foundations and boring plumbing. It doesn’t define your success metrics.
3. Invest in literacy and experimentation.
Your goal is to move people to “consciously incompetent” about AI. They should know enough to say, “I think we could automate this,” even if they don’t know how.
4. Design for augmentation, not replacement.
Repeat Glen’s line to your teams: automation is about “taking the robot out of the human”. Roles will change. They should also become more interesting.
5. Get something useful in front of people before the data is perfect.
Let your users help you discover where the real data problems are. Just don’t treat them as unpaid QA – respect their time and close the loop when they flag issues.
6. Keep humans accountable.
No “the AI did it” excuses. If your name is on the presentation, email or escalation, you own it. AI is a tool, not a shield.
7. Make AI everyone’s job.
Support from central teams is important. But the real responsibility sits with the people who own outcomes in Sales, Marketing, Product, Finance and Operations.
If you want to hear Glen in his own words – including his story about working with the creators of R back in the 90s, why he loves Otter, and why he thinks we’ll still be fighting about hype five years from now – you can listen to the full episode of Building with AI: Promises and Heartbreaks with him
And if you’re trying to make sense of where AI actually fits in your own data and analytics journey, we’re always happy to talk about how Solid can help.