Key takeaways
- The current AI gap is not access; it is operating impact. Stanford HAI reports broad 2025 organizational AI adoption, while McKinsey finds only a small high-performer group reporting significant value and 5%+ EBIT impact.
- Middle market companies can compete by being disciplined: one workflow, one owner, one baseline, one review standard, then expand.
- The strongest AI business cases start with work that is already recurring, measurable, and reviewable: management reporting, diligence response, procurement research, and close support.
- A company that cannot show before-and-after metrics is not implementing AI; it is experimenting with AI.
Stanford HAI reports that surveyed organizational AI use reached 88% in 2025 and regular generative AI use reached 79%, making AI access a common condition rather than a durable advantage.
McKinsey defines AI high performers as organizations reporting significant value and at least 5% EBIT impact from AI; that group represented about 6% of survey respondents, which highlights the gap between use and operating results.
Federal Reserve analysis of Census and other surveys shows that U.S. AI adoption measurements vary based on whether the question captures production use, any business-function use, worker-level GenAI use, or employment-weighted firm adoption.
88%
Surveyed organizations using AI in at least one function in 2025
79%
Surveyed organizations regularly using GenAI in at least one function
6%
McKinsey AI high performers reporting significant value and 5%+ EBIT impact
The AI conversation has moved past the novelty phase. A middle market company no longer earns operational credit because employees have access to ChatGPT, Copilot, Claude, or embedded AI features in existing software. That access is now table stakes. The question that matters is whether AI has changed how the business runs: fewer hours spent assembling management reports, faster diligence responses, cleaner procurement analysis, more consistent close support, and better operating visibility.
The execution gap appears when AI is present but not institutionalized. Staff use tools opportunistically. A few prompts circulate in Slack. One finance manager saves time on commentary, while another still starts from a blank page. No one has a baseline, no one tracks quality, and no one can say which workflows have actually improved. The business is using AI, but the operating system has not changed.
What separates adoption from execution
AI adoption means a tool is available or used. AI execution means a defined workflow runs differently because of it. The difference is not semantic; it is the difference between convenience and operating leverage.
The companies that move from adoption to execution make the workflow the unit of analysis. They do not ask, "How should we use AI?" They ask, "Which recurring workflow consumes measurable time, has a definable output, and is owned by one person who can review and improve the result?" That framing turns AI from a technology conversation into an operating improvement conversation.
The middle market advantage
Large enterprises have more resources, but they also have more complexity: fragmented data architecture, longer approval cycles, multiple risk teams, and broader change-management requirements. Middle market companies can move faster if they choose narrower use cases and keep governance practical.
The advantage is focus. A $20M business does not need an enterprise AI transformation office to create value. It needs a controller-owned management reporting workflow, an operations-owned vendor research workflow, a COO-owned meeting action workflow, or a CFO-owned diligence response workflow. Each one should have a baseline, an output standard, a review process, and a measured result.
The most valuable AI implementation in a middle market company is usually not the most advanced one. It is the first workflow that becomes boring, reliable, measured, and part of the operating cadence.
How to close the gap in 90 days
The first 90 days should be narrow. Pick one workflow, measure the current state, write the output standard, assign one owner, run the workflow through AI assistance for three to five cycles, and report the result. If the result is measurable, expand to the next adjacent workflow. If the result is not measurable, fix the workflow design before expanding.
90-Day AI Execution Plan
The discipline is intentionally simple. AI execution does not begin with an enterprise platform search. It begins with a workflow where the business can prove, in dollars, hours, quality, or speed, that the operating system improved.
Work with Glacier Lake Partners
Map Your AI Execution Gap
Identify where AI adoption exists but measurable operating impact has not yet been captured.
Request an AI Scan →Research sources
Disclaimer: Financial figures and case studies in this article are illustrative, based on representative middle market assumptions, and are not guarantees of outcome. Statistical references are drawn from cited third-party research; individual transaction and operational results vary based on business characteristics, market conditions, and deal structure. This content is for informational purposes only and does not constitute legal, financial, or investment advice. Consult qualified advisors for guidance specific to your situation.

