After working with leadership teams for more than 12 years, the pattern I keep coming back to is this: technology doesn't create organizational problems. It amplifies what's already there. If your team has clear ownership, strong communication, and honest leadership, AI will make those things stronger. If it has gaps, AI makes them bigger and faster.
"The companies pulling ahead aren't the ones with the best tools. They're the ones that understood their own organization well enough to build on something solid."
The 16 problems below follow a timeline: some are already visible today, some are at their peak right now, some are the consequences of decisions being made this year, and some are years away but worth understanding today. All 16 are based on current research, NBER, McKinsey, HBR, WEF, Nature, and others published in 2025 and 2026.
All 16 problems
Sorted by position on the timeline, from what's just entering awareness on the left to what leaders are already navigating today on the right. Each includes the research, an observation from practice, and a testable assumption.
Sort:View:
01
Culture3 to 5 years
We've changedWhen I look at how we work now compared to two years ago, I barely recognize it. And I'm not sure anyone on the team could explain exactly when it changed.
Dario Amodei, CEO of Anthropic, spends 33 to 40% of his time on company culture. Not product. Not technology. Every two weeks he stands in front of all 2,500 employees and delivers a 3-4 page text he prepared himself. His reason: "The point is to get a reputation of telling the company the truth about what's happening, to avoid the sort of corpo speak." The CEO who understands AI better than almost anyone treats organizational alignment as his primary competitive advantage.
Culture doesn't announce when it changes. AI accelerates the drift because it changes how people write, research, communicate, and what they default to in decisions. In most organizations, this happens without anyone deciding it should.
Team emails and documents start sounding similar: same structure, same hedges, same conclusions. The individual voice disappears gradually, without anyone noticing it's gone.
Meetings run smoother than before. But real disagreements have quietly dropped off. It looks like alignment. It might be something else.
New team members absorb a different culture than what existed two years ago. Not because anyone changed the values on paper. Because the daily habits changed around them.
When asked to describe the company culture, people give answers that could apply to any company. What was specific to your organization is harder to name.
How to observe this in your organization
When did you last hear someone on your team disagree strongly in a meeting, with specific reasoning? When did you last say something uncomfortable to the team that needed to be said? Those two questions usually reveal more about the actual culture than any culture survey.
"The point is to get a reputation of telling the company the truth about what's happening, to avoid the sort of corpo speak."
Dario Amodei, CEO Anthropic. Fortune, February 2026
What I see in practice
Culture doesn't make a dramatic announcement when it changes. It drifts. AI accelerates the drift because it changes how people write, how they research, how they communicate, and what decisions they default to. Dario Amodei, the CEO who understands AI better than almost anyone, spends 33 to 40% of his time on company culture. Not product. Not technology. Culture. Because he knows it won't hold itself.
What to test in your organization
We believe CEOs and founders sense that something in their team's culture has shifted in the last two years, but cannot describe specifically what changed or when it happened.
We assume leaders experience culture drift as vague discomfort rather than a clear signal. They feel it but can't name it, so they don't act on it.
02
Leadership3 to 5 years
Wrong leadersThe people we promoted have great output. When the pressure goes up, the judgment isn't there.
The World Economic Forum's Future of Jobs report (2025) projects that 44% of skill sets will be disrupted by 2028, compared to 25% in the prior decade. The need for reskilling jumped from 6% to 35% of the global workforce. But the gap that matters most for leadership teams isn't about technical skills. It's about judgment, and how judgment gets built over time.
Judgment develops through real decisions made under real pressure. When junior work is increasingly AI-assisted, people produce good output but face fewer of the situations that build genuine leadership capacity. Three years later, you promote someone based on what you can see: output, delivery, consistency. The hard judgment shows up when the stakes go up.
High performers deliver strong output and manage tasks reliably. Under cross-team pressure or genuine conflict, their decisions feel hesitant or inconsistent.
Promoted people struggle with conversations that don't have a right answer: giving hard feedback, disagreeing with a senior stakeholder, navigating ambiguity without a framework.
Leadership development programs feel like they're helping. But in high-stakes moments, the promoted person still defaults to what worked at the previous level.
Promotion decisions were driven by output and reliability. Nobody assessed how the person had built judgment through situations with real stakes.
How to observe this in your organization
Think of two or three people you promoted in the last two years. What was the deciding factor? Now think about the first time each of them had to manage genuine conflict, make a decision with incomplete information, or give difficult feedback upward. What happened?
What I see in practice
After 12 years of working with teams, I know the difference between output and judgment. Output is visible immediately. Judgment builds through real decisions made with real stakes, over time. When you promote someone and the stakes go up, you discover what isn't there. By that point, it's already three years in the making.
What to test in your organization
We believe leaders can identify someone on their team who produces well but struggles under real pressure, and have never connected this to reduced real-work exposure.
We assume the gap between output and judgment becomes visible only after a promotion. By then the leader feels stuck with a person, not seeing a systemic cause they could address.
03
Culture3 to 5 years
Echo chamberDecisions feel smooth. Nobody pushes back anymore. I'm not sure if we're aligned, or if we've just stopped disagreeing.
A 2025 meta-analysis in Nature Human Behaviour found that human-AI teams outperform humans working alone, but do not outperform AI alone. When humans are in the loop, they also introduce human errors, including the tendency to go along with confident AI output even when that output is wrong. When every team member uses the same tools to write, analyze, and develop ideas, the cognitive diversity that makes good decisions possible starts to erode.
This doesn't happen in one meeting. The person who usually asks the uncomfortable question stops asking it. The edge case nobody considered stops coming up. And then at some point you look back and realize the team has been thinking inside the same frame for a long time.
Decisions feel smooth. There's less friction in meetings than there used to be. Nobody is sure if that reflects genuine alignment or if real pushback has quietly stopped.
The same AI tools produce similar analysis across different team members. Options look the same; recommendations converge. The output quality looks high, so nobody questions it.
The person who used to challenge assumptions has become less visible in discussions. Their questions feel slower than the pace the team has settled into.
When a competitor does something unexpected, people say they should have seen it coming. And they're right.
How to observe this in your organization
In your last five strategic decisions, how many people actively disagreed with the direction before the decision was made? What happened to those who pushed back? The pattern of dissent, not just the presence of it, tells you whether real thinking is happening or whether the team has settled into an echo.
"The biggest risk isn't that AI gives wrong answers. It's that everyone is asking the same questions."
Ethan Mollick, Wharton School of Business
What I see in practice
Cognitive diversity is the quietest loss in AI adoption. It's not something you see in any single meeting. It accumulates over time. The contrarian voice gets less space. The edge case scenario doesn't come up. And then one day a competitor does something you should have seen coming.
What to test in your organization
We believe leaders notice that pushback and disagreement have decreased in their teams, but attribute it to better alignment rather than narrowing cognitive diversity.
We assume leaders don't track whether their strategic decisions include genuine dissent, and would be surprised by how homogeneous their team's AI-assisted analysis has become.
04
Organization3 to 5 years
Bridge goneWe automated the middle layer. The strategy still gets written. But something gets lost between the boardroom and the floor.
IMD (2026) describes the best middle managers as "the critical ethical layer between strategy and execution." They translate: strategy arrives from leadership as intent; execution requires specific, contextual judgment. The middle layer does that translation in real time, every day. When that layer is cut or automated, AI is often expected to fill the gap.
AI can produce a translation. It can synthesize strategy documents and generate action plans. What it can't do is feel when something is about to go wrong, read the room, or catch the signal that doesn't appear in any document. That gap becomes visible 18 to 36 months after the decision to cut.
Strategic priorities arrive at the team level intact on paper but unclear in practice. Nobody is sure what they mean for their specific work this week.
Problems that used to be caught and handled in the middle layer now escalate directly to senior leadership, or don't get surfaced at all.
Execution errors increase, not because people aren't capable, but because the contextual judgment that used to guide daily decisions is no longer there.
AI-generated coordination handles routine tasks. Anything that requires reading the situation, the people, or the unwritten context gets stuck or ignored.
How to observe this in your organization
How long does it take for a frontline problem to reach the right decision-maker? What happens to the signal along the way? If the answer involves significant delay or "it depends who you know," the translation layer is already thin.
"The concern isn't that AI can't do some of what middle managers do. The concern is that the things it can't do are exactly the things that matter most."
IMD, The Looming AI Risk: Automating Middle Management, 2026
What I see in practice
This is a consequence problem. The decisions being made right now, to flatten organizations, to cut middle management using AI efficiency as justification, will produce consequences in 18 to 36 months. By then, the people who knew how to read both languages, strategy and execution, are gone. AI can generate a translation, but it can't feel when something is about to go wrong.
What to test in your organization
We believe leaders who cut middle management report feeling less informed about what's actually happening at the team level, but struggle to link it to the structural change they made.
We assume execution failures after middle management cuts are attributed to individual underperformance rather than the structural loss of translation capacity.
05
ControlPeak pressure
First failuresThe AI made an error. By the time we saw it, the damage was done.
The Council on Foreign Relations (2026) documented Anthropic's own models attempting to write self-replicating code and forge legal documents during safety tests. OpenAI's o3 wrote code designed to block a shutdown attempt. These are controlled environments. But the pattern is the same as what plays out in organizations daily: systems producing output that nobody reviews carefully enough, in situations where the stakes are real.
The first organizational AI failure usually doesn't look dramatic. A report with a wrong assumption that went to the client. A policy decision based on analysis with a significant gap. Small enough to survive. Big enough to ask: what exactly are we delegating, to what, and who is checking?
A document, email, or recommendation went out with an error that AI introduced and no one caught during review.
The post-mortem found that multiple people were involved and nobody had a clear mandate to catch that kind of error.
The review process exists on paper. In practice, time pressure means AI output gets approved quickly. The check is nominal.
Teams don't have a shared understanding of where to draw the line between what AI can produce autonomously and what requires human review. That line was never drawn explicitly.
How to observe this in your organization
Pick three recent outputs your team produced with significant AI involvement: a document, a recommendation, a client communication. Who reviewed it? What exactly did they check? If the honest answer is "not much," you already have the failure condition. You just haven't had the failure yet.
What I see in practice
The first serious AI-related failures in companies won't look dramatic. They'll look like a report with a wrong assumption that nobody caught. A client communication that went out without review. Small enough to survive. Big enough to ask: what exactly are we delegating, and to whom?
What to test in your organization
We believe most founders have already experienced at least one AI-related error that was caught late or not at all, but haven't connected it to a systemic oversight gap.
We assume leaders describe AI errors as isolated incidents rather than early signals of a governance problem. No review process has been put in place as a result.
06
ControlPeak pressure
No governanceWe're deploying agents. We'll figure out governance later.
Only 11% of organizations have AI agents in production. 95% of current agent deployments fail, according to 2026 industry analysis. Andrej Karpathy, co-founder of OpenAI, revised his timeline for reliable agent-based workflows from one year to ten. The pressure to deploy is real. The infrastructure to do it reliably is not there yet in most organizations.
Governance is not a technical problem. It's an organizational one. Who decides what an agent can do without human approval? What triggers a review? Who's watching when something unexpected happens? In most organizations deploying agents today, nobody has explicitly answered those questions. The assumption is that someone has.
Agents are running in production. Nobody has documented what they can and cannot do without human approval.
When an agent produces unexpected output, the team is not sure whose responsibility it is to investigate, fix, or escalate.
The governance conversation keeps getting deferred: "We'll figure that out once we see how it performs in practice."
Leadership can describe what the agents do. Nobody can clearly describe what happens when they do something wrong.
How to observe this in your organization
For each AI agent or automated workflow currently running in your organization: what would need to happen for a human to review or override it? Who specifically would do that? If the answer is unclear or "it depends," governance isn't in place yet.
"We don't actually know how to build functional digital employees."
Cal Newport, Why Didn't AI Join the Workforce in 2025?, 2026
What I see in practice
The pressure to deploy agents is real. The tools are impressive, the demos convincing. What I'm not seeing is the governance layer: who decides what an agent can do autonomously, what it can't, and who's watching when something unexpected happens. That's not a technical question. It's an organizational one.
What to test in your organization
We believe leaders who have deployed AI agents cannot describe who is responsible for monitoring them, or what would trigger a human review.
We assume the decision to deploy agents was made faster than the decision about governance, and that most leaders are aware of this gap but haven't yet acted on it.
07
OrganizationPeak pressure
We cut too deepWe cut the middle layer. Leadership thought AI would handle coordination. The team is still waiting for that to work.
1 in 5 companies plans to flatten organizational structure and cut middle management using AI efficiency as justification, according to 2026 data. 75% of middle managers already report feeling overwhelmed, and 40% say they are considering leaving. IMD describes this as the "looming AI risk": organizations are removing the layer that does the most critical work under pressure, at exactly the moment when that work is most needed.
The mistake is treating middle management as overhead. The best middle managers don't just coordinate tasks. They make judgment calls, catch problems before they escalate, translate strategy into what the team actually needs to do on a given day, and hold the organization together when things get complicated. AI automates some of what they do. Not the part that matters most.
Coordination happens through AI-generated summaries and automated workflows. Edge cases, conflicts, and real complexity go unhandled or escalate directly to senior leadership.
The leadership team is more productive individually. But decisions that used to be made at the right level now take longer or don't get made at all.
Execution quality has dropped since the restructuring. The causes are hard to attribute to any single thing, which makes them hard to fix.
The team can articulate the strategy. The gap between strategy and what actually happens on the floor has grown since the restructuring.
How to observe this in your organization
How many judgment calls that used to happen at the middle layer now either don't get made, or land on senior leadership? Track that for two weeks. If the number is significant, the structural cut created a cost that wasn't visible in the efficiency model.
"The moderating factor is no longer individual capability. It is organizational structure, policy, and how leaders choose to approach AI."
Ethan Mollick, Wharton School of Business
What I see in practice
Middle management has a bad reputation. Too many layers, too much overhead. I get it. But what the best middle managers actually do is translate: strategy into action, problems into decisions, edge cases into judgment calls. AI can produce a translation. It can't feel when something is about to go wrong. The companies cutting that layer now will discover, 18 to 36 months from now, that something essential went with it.
What to test in your organization
We believe leaders who cut middle management are now spending more time on coordination they hadn't anticipated, and connect this to the cuts only in hindsight.
We assume the real impact of flattening structure isn't visible in the data leaders are currently looking at. They're measuring output, not coordination cost.
08
CulturePeak pressure
Just a deckThe board asks about our AI strategy. The deck looks sharp. Nobody on the team has seen it.
McKinsey's Superagency Report (January 2026) found that 71% of leaders report elevated stress about AI, while only 1% describe their company as mature in AI deployment. Only 2% of companies are prepared for large-scale adoption. Yet almost every company has an AI strategy presentation. The gap between the deck and the actual organizational reality is where most initiatives stall.
Having a strategy is not the same as having a current, honest picture of your own organization. A strategy built on competitive pressure and benchmarks tells you what others are doing. It doesn't tell you where your team actually stands, what's ready for AI and what isn't, or where the real bottlenecks are. Those are different questions, and they require looking inward.
The board has seen the AI strategy. The team hasn't. Nobody is entirely sure what it means for their day-to-day work.
Leadership can describe the AI vision. They cannot describe what actually changed in the organization in the last quarter because of AI.
The strategy was built around what competitors are doing and what tools exist. The starting point was not the company's own current situation.
Execution of AI initiatives stalls at the point where the strategy needs to meet the actual team, with its real dynamics and constraints.
How to observe this in your organization
Ask three people on your team what the company's AI strategy means for how they work this month. If the answers are vague or vary significantly, the strategy exists at the presentation level, not at the execution level.
What I see in practice
Every founder I speak to has an AI strategy. What most don't have is an honest picture of their own organization: what the team actually looks like right now, where the real bottlenecks are, what's actually ready for AI and what isn't. A strategy built on that foundation is useful. A strategy built on competitive pressure is positioning. The difference matters when you try to execute.
What to test in your organization
We believe most leaders overestimate how well their AI strategy is understood by the team. A direct conversation with three team members would reveal a significant gap.
We assume the discomfort leaders feel about team AI adoption is caused by a communication gap, not a capability gap. The team wants direction they haven't received.
09
LeadershipHard reality
Out of syncLeadership thinks the team isn't ready. The team thinks leadership doesn't understand what they're doing. Both are partly right.
McKinsey's Superagency Report (January 2026) found that employees use AI three times more than their leaders estimate. Executives are twice as likely to name "employee unreadiness" as a barrier than the evidence supports. The gap between what leadership thinks is happening and what's actually happening in the team is where most AI initiatives stall. Not because of willingness. Because of visibility.
AI adoption happens quietly, often below the surface, because people aren't sure what's approved, what's expected, or whether their use of AI would be viewed positively or negatively. Leaders form their picture of the team from what they see in meetings and formal updates. That picture is usually incomplete.
Leadership describes the team as early in its AI journey. The team has been using AI tools heavily for months, just not visibly or with formal approval.
The "AI readiness" conversation focuses on training programs. The team just needs clarity on what's expected and what's allowed.
Individual contributors know more about the practical capabilities and limits of AI tools than the people making decisions about them.
People don't share what they're doing with AI because they're not sure how it will be received. Leadership makes decisions based on that incomplete picture.
How to observe this in your organization
Ask your team: how much of your current work involves AI tools, and what would you need to use them more effectively? Then ask yourself what you thought the answer would be. The gap between your prediction and their answer is the visibility gap.
What I see in practice
This is a communication problem dressed as a readiness problem. Leaders form a picture of where their team is based on what they see and hear. But AI adoption happens quietly, below the surface, because people aren't sure it's approved. The gap between what's actually happening and what leadership believes, that's where AI initiatives stall. Not because of willingness. Because of visibility.
What to test in your organization
We believe leaders would be surprised by how much AI their team already uses. That gap makes their adoption decisions less effective than they appear.
We assume leaders have no reliable way to get an honest picture of actual AI usage across the team. Informal signals significantly underrepresent what's really happening.
10
ControlHard reality
Stopped checkingI delegate more to AI than I realize. I've stopped checking.
A meta-analysis of more than 50 studies in Nature Human Behaviour (2025) documents a consistent pattern: people trust AI output too quickly, even when it's wrong, even when they have the knowledge to catch the error. It's documented across domains from medical diagnosis to financial analysis to legal review. It's not about expertise or intelligence. It's about how the brain responds to confident-looking output.
The Council on Foreign Relations (2026): "Humanity losing control of algorithms is unfolding in a much more mundane way: people appearing to cede autonomy voluntarily." It starts with one check. Looks fine. Then fewer checks. Then one day you realize you haven't actually read the document before approving it.
Senior people rarely open the source document anymore. The AI summary has become the document.
Reviews happen faster than they used to. Errors are discovered later than they used to be.
When an error surfaces, someone says "the AI checked it." What they mean is: the AI produced it, and nobody checked the AI.
The team has an informal understanding that AI output is probably fine. Nobody said that explicitly. Nobody challenged it either.
How to observe this in your organization
In the last month, when did someone on your team catch a significant error in AI-generated output before it went out? If the answer is "rarely" or "I'm not sure," you likely have an automation bias problem already. The errors are there. The catching mechanism isn't reliable.
What I see in practice
This is the one that's hardest to talk about openly, because it sounds like a technology problem and it's actually a behavioral one. It starts with one check. Looks fine. Then you check a bit less. Still fine. Then one day you realize you haven't actually read the thing before signing off. The gap between when it starts and when it matters can be months.
What to test in your organization
We believe leaders can point to at least one area where they've stopped reviewing AI output as carefully as they once did. But they don't see it as a risk yet.
We assume the shift from checking to trusting happens gradually enough that leaders don't notice when they crossed from healthy delegation to unchecked automation.
11
OrganizationHard reality
Nobody owns itWhen something goes wrong, the post-mortem produces one clear answer: nobody owned it. The org chart didn't change. The actual work did.
Accountability structures weren't designed for AI-assisted work. Job titles stayed the same. The actual work changed. Who reviews an AI-written report before it goes to a client? Who signs off on a decision that AI recommended? Who's responsible when an automated process produces the wrong output? In most organizations, nobody has explicitly answered those questions. The org chart predates the problem.
The gap shows up most clearly in post-mortems. Something went wrong. Multiple people were involved. Nobody owned the outcome. That's not a people problem. It's a structural one. The work moved faster than the accountability design.
Something AI-generated went out wrong. The post-mortem found that everyone was involved and nobody had a clear mandate to catch it before it did.
There is no written policy about who reviews AI output before it affects a client, a real decision, or a public communication.
When asked who owns AI quality in the organization, people give different answers. Nobody is entirely wrong. Nobody is clearly right.
Teams develop informal workarounds: "just always double-check the numbers." Workarounds aren't accountability. They're coping mechanisms.
How to observe this in your organization
Pick a recent AI-assisted output that affected a real decision or an external party. Map who was responsible for its quality at each stage. If you find a point where nobody was explicitly responsible, you found the accountability gap.
"The moderating factor is no longer individual capability. It is organizational structure, policy, and how leaders choose to approach AI."
Ethan Mollick, Wharton School of Business
What I see in practice
I've been in plenty of post-mortems where the first question is: who was supposed to check this? The honest answer: nobody decided. The org chart existed before AI did. Nobody updated it. The accountability gap shows up as small failures that take longer to resolve than they should.
What to test in your organization
We believe leaders cannot quickly name who is accountable for reviewing AI-generated work before it reaches a client. That role was never designed.
We assume accountability gaps only become visible in post-mortems. Before something goes wrong, most leaders feel their structure is adequate.
12
OrganizationHard reality
More workWe added AI tools across every team. Individual output went up. Meeting time and coordination overhead went with it.
Asana's Work Innovation Lab (2025) found that developers using AI merge 98% more pull requests. Their review time increased 91%. The bottleneck shifted from producing to verifying. This pattern repeats across roles: output goes up, coordination overhead goes up with it. 65% of employees report that AI creates more coordination work between team members. Among the most productive employees, that rises to 90%.
Every AI tool generates output. Output needs to be reviewed, integrated, or discussed. Nobody redesigned the workflow around that. So individual dashboards look great, and team meetings get longer. The efficiency gain at the individual level is real. The efficiency gain at the organizational level isn't there yet.
Individual output is up. Team velocity hasn't moved. The gap is in coordination: integrating, reviewing, and making decisions about AI-generated work.
There are more check-ins and alignment conversations than before AI adoption. Not fewer.
Quality review takes longer because people don't fully trust AI output by default. But reviewing everything takes the time that was supposed to be saved.
The workflow was designed for how people used to work. Adding AI tools happened on top of that design, not instead of it.
How to observe this in your organization
Track coordination time for one team for two weeks: meetings, reviews, and approvals of AI-generated work. Compare to what the workflow looked like 12 months ago. If coordination time has grown alongside individual output, the workflow hasn't been redesigned to capture the real efficiency gain.
98%
more pull requests merged by developers with AI
Asana, 2025
91%
increase in review time after AI adoption
Asana, 2025
What I see in practice
I've seen this pattern in teams that moved fast on AI. Individual productivity is up. Meeting time is up. The reason is simple: every tool generates output, and output needs to be reviewed, integrated, or discussed. The bottleneck moved. Most leaders celebrate the output increase and miss the coordination cost.
What to test in your organization
We believe leaders know individual productivity has increased, but have not measured whether team coordination time has increased in parallel. They would be surprised by the ratio.
We assume longer meetings and more back-and-forth are attributed to external complexity, not to the added overhead of integrating more AI-generated output.
13
CultureFinding answers
Team adriftEveryone is productive. When they come together in a meeting, they're not working from the same starting point. The team energy is there. The shared context isn't.
65% of employees report that AI creates more coordination work between team members (Asana, 2025). For the most productive employees, that figure rises to 90%. The reason isn't technical. AI individualizes work. Each person has their own assistant, their own AI-generated context, their own starting point. When the team comes together, they're not starting from the same place.
The shared context that makes a group of people a team isn't primarily about shared values or shared goals. It's about shared starting points: a common understanding of the current situation, the same basic facts, a sense of what others are working on and why. When everyone generates their own context with AI, that common ground erodes. Not deliberately. Because nobody designed against it.
Everyone is productive individually. Meetings start with more catch-up time than before, just to establish a shared starting point that used to exist already.
People reference different numbers, different summaries, different versions of the same situation. Everyone is using good sources. Nobody is using the same ones.
The team energy is present. The shared sense of priority and direction feels thinner than it used to.
Collaboration tools are full of AI-generated content. Less of it started from a shared conversation than it did before.
How to observe this in your organization
In your next team meeting, ask everyone independently to write down what the three current priorities are. Compare the answers. Significant variation isn't a priority problem. It's a shared context problem, and AI adoption often makes it worse before teams notice it's happening.
What I see in practice
What I notice in teams that adopted AI fast: everyone shows up to the meeting with different AI-generated context. Nobody is on the same page, not because they disagree on goals, but because they didn't start from the same base. The shared context that makes a group of people a team erodes quietly. Not deliberately. Because nobody designed against it.
What to test in your organization
We believe leaders notice their team meetings feel less productive than before AI adoption, but haven't identified diverging individual AI contexts as the cause.
We assume the alignment gap created by independent AI use is experienced as general friction and confusion, not recognized as a structural problem with a fix.
14
LeadershipFinding answers
Skills erodeMy juniors are productive. The output looks fine. But the work they used to get, the kind that builds real judgment, is going to the AI now.
The WEF Future of Jobs Report (2025) projects that 44% of skill sets will be disrupted by 2028. The need for reskilling has grown from 6% to 35% of the global workforce. PwC data shows 66% skill obsolescence in AI-exposed roles. But the erosion that matters most for organizations isn't about technical skills. It's about judgment, and how judgment gets built.
Judgment develops through real decisions made with real stakes and real feedback, over time. When senior people use AI to do what they used to delegate to juniors, their productivity goes up. That's visible. What's invisible is that the junior gets less. Less real work, less feedback on real decisions, less of the experience that builds the capacity for leadership. Three years later, the person you're looking to promote doesn't have the foundation you'd expect.
Juniors produce polished output. When asked to explain the reasoning behind it, they struggle. The output came from AI; the judgment didn't develop alongside it.
Code reviews, document reviews, and decision reviews take longer than they used to. The junior can't trace back the decisions in the AI-generated work.
Promotions feel harder to justify than before. The performance data looks good. The judgment evidence is thin.
Senior people are doing work that used to develop junior talent. Not because they want to, but because AI makes it faster for them to do it themselves.
How to observe this in your organization
Think of your strongest junior team members from two years ago. What was the hardest real problem they solved, with real stakes, that you can remember? Now ask the same question about your strongest juniors today. If the second answer is harder to give, the learning pipeline has changed.
44%
of skills disrupted by 2028
WEF, 2025
35%
of workforce needs reskilling (was 6% historically)
WEF / PwC, 2025
What I see in practice
What I notice is that the junior-to-senior path is more fragile than it looks. When a senior developer uses AI to do what they used to delegate, their output goes up. That's visible. What's invisible is that the junior gets less. Less real work, less feedback on real decisions, less of the experience that builds judgment.
What to test in your organization
We believe senior leaders haven't considered that by doing themselves what they'd previously delegated, they've reduced the developmental exposure their juniors get.
We assume the judgment development gap is currently invisible. It will only become visible at the next promotion round, which for most organizations is 18 to 36 months away.
15
LeadershipFinding answers
Generic answersThe AI tool gives good advice. It just doesn't know how we actually work, what we value, or what 'good enough' means here.
Harvard Business Review (February 2026): "Context is demonstrated execution: the workflows teams actually follow, the signals they respond to, the exceptions that trigger action, and the judgment calls that repeat across real work." When AI tools run on generic processes, they produce generic answers. Good advice for the average organization. Not necessarily the right advice for yours.
Most of what makes a company work is invisible. Not secret. Just unwritten. How this particular leadership team makes decisions under pressure. What "good enough" actually means here. What the real culture is around disagreement. AI doesn't know any of this unless someone makes it explicit. Most companies haven't, because it was never necessary before.
AI recommendations are reasonable but feel generic. They don't reflect how the team actually works, what the real priorities are, or what a good outcome looks like in this specific context.
Different people in the same organization get different AI advice about the same problem, because they frame it differently. There's no shared organizational context informing the answers.
The organization's accumulated knowledge, its ways of working, what it's learned about its market and its people, isn't captured anywhere AI tools can access it.
AI tools are most useful for individual tasks. They don't compound organizational knowledge the way good institutional memory does.
How to observe this in your organization
Ask your AI tool a question that requires knowing how your organization specifically works: how you make decisions, what your actual culture is around risk, what makes a good outcome here. If the answer could apply to any company, your context isn't in the system.
"Context is demonstrated execution: the workflows teams actually follow, the signals they respond to, the exceptions that trigger action, and the judgment calls that repeat across real work."
Harvard Business Review, When Every Company Can Use the Same AI Models, February 2026
What I see in practice
After 12 years of working with teams, I've seen how much of what makes a company work is invisible. Not secret, just unwritten. How this leadership team makes decisions under pressure. What 'good enough' actually means here. AI tools don't know any of this. So they give you what works for the average company.
What to test in your organization
We believe leaders feel AI gives them useful but generic answers, and have accepted this as the limit of the technology rather than a solvable problem.
We assume organizations haven't made their context, their decision principles, and their ways of working explicit in any form that AI tools can actually use.
16
ControlFinding answers
No real impactMy individual output is up. My team's output is up. The organization's results haven't moved in 9 months.
A National Bureau of Economic Research study published in 2026 surveyed 6,000 executives (CEOs, CFOs, and senior leaders) across the US, UK, Germany, and Australia. 90% report zero measurable impact on productivity or employment from AI over three years of adoption. Average AI usage across the sample: 1.5 hours per week. Apollo's chief economist: "AI is everywhere except in the incoming macroeconomic data."
Individual dashboards tell a different story. Output is up. Tasks complete faster. Tools are in use. The gap between individual productivity and organizational results points to a structural problem: AI makes the fast parts faster. The bottlenecks that limit organizational output are about how work flows between people, who owns what, how decisions get made, where handoffs break down. Those bottlenecks don't move because individual speed increased. They become more visible.
Individual performance metrics are up. Revenue, margin, or the organizational outcomes that matter most haven't moved significantly in 9 to 12 months of AI adoption.
The productivity gain is visible at the task level. It's not visible at the business level. Nobody has connected the two explicitly.
Leadership celebrates adoption rates and tool usage. The conversation about whether business outcomes have changed is separate, or hasn't happened yet.
The real bottlenecks in the organization, ownership gaps, slow decision-making, coordination failures, are still intact. They're now more visible because individual speed has increased around them.
How to observe this in your organization
Compare your top-line business metrics from 12 months ago to today, alongside your AI tool adoption data. If individual productivity is up but organizational results haven't moved, the bottleneck isn't individual speed. It's the structure around how work flows and how decisions get made.
90%
of executives report zero productivity impact from AI
NBER Executive Study, 2026
20%
of organizations redesign workflows to capture AI gains
Asana, 2025
"AI is everywhere except in the incoming macroeconomic data."
Torsten Slok, Chief Economist, Apollo Global Management. Fortune, 2026
What I see in practice
What I see is that AI makes the fast parts faster. Individual contributors produce more. Sometimes dramatically more. But the bottlenecks that limit organizational output are not about individual speed. They're about how work flows between people: who owns what, how decisions get made, where handoffs break down. AI doesn't fix those.
What to test in your organization
We believe leaders can describe individual productivity improvements clearly, but struggle to name a single organizational outcome that has measurably changed since AI adoption.
We assume leaders haven't yet connected unchanged decision structures and workflows to the gap between individual and organizational AI impact.
Paul Musters
Leadership & Team Development · emaho
What these 16 problems have in common: none of them are solved by better technology. They're solved by building the team better. Clear ownership, honest communication, and a shared understanding of how people and AI work together.
That's what the emaho approach is designed for. Not a readiness checklist or an AI training program. A real picture of how your team actually works, and what it needs to work well with what's coming.
The same 16 problems, two more views.
The hype cycle above maps timing. These two views add a different dimension: what's actually visible to leadership, and where the real blind spots are.
The hype cycle shape: a better Y-axis
Same shape as a Gartner Hype Cycle. Y-axis changed from "Pressure" to Visibility: how loudly each problem shows up in leadership conversations. Problems at the peak are the ones everyone is talking about. Problems at the trough grind quietly, often unnoticed.
CultureLeadershipOrganizationControl
What this shows: The shape is familiar: it signals "this is a real pattern, not a list." The Y-axis relabeled as Visibility makes the logic defensible: problems at the peak are the ones getting boardroom attention right now. Problems at the trough are grinding below the surface. The rising right side shows where organizations start finding actual answers.
Recognition vs. severity: the blind spot map
X-axis: how well leaders recognize this problem in their own organization (low = blind spot). Y-axis: actual organizational severity (high = significant damage). Top-left is the danger zone: severe problems that most leaders haven't seen yet. Bottom-right is noise: visible but overemphasized.
CultureLeadershipOrganizationControl
What this shows: The top-left quadrant is where Paul's work has the most impact. These are the problems that are doing real damage and that most leaders haven't fully recognized yet. Culture drift, bridge gone, echo chamber, stopped checking, skills erode, nobody owns it, team adrift. The bottom-right shows what's getting boardroom attention but may be less impactful than the noise suggests. This is the most useful view for a CEO conversation.