The AI Application Layer Is Finally Working
After two years of skepticism about AI application companies, we are seeing clear evidence that the application layer is working. The best AI applications combine three things: a s...
Model provider dynamics directly affect enterprise AI strategy. The current price collapse and capability convergence are creating a window where enterprises can negotiate favourable terms and avoid lock-in.
The intensifying competition among frontier model providers including OpenAI, Anthropic, Google DeepMind, Meta, and Mistral. Covers capability benchmarks, pricing wars, and enterprise licensing models.
| Source | Type | Items |
|---|---|---|
| Sequoia Capital | Vc pe | 2 |
| a16z Podcast | Podcast | 1 |
| Acquired | Podcast | 1 |
| Stratechery | 1 | |
| The Information | 1 | |
| Bessemer Cloud Index | Vc pe | 1 |
| @benedictevans | X influencer | 1 |
| @saboreman | X influencer | 1 |
| Lex Fridman Podcast | Podcast | 1 |
AI agents are the new aggregators. Just as Google aggregated web content and Facebook aggregated social connections, AI agents will aggregate services and workflows. The company that owns the agent layer -- the interface between the user and all the services they use -- captures the majority of the value. This is why every platform company is racing to be the default agent.
Act One of generative AI was about model capabilities and the infrastructure to serve them. Act Two is about applications that generate sustainable revenue. We are now seeing the first generation of AI-native applications reach $100M ARR, and they share common traits: they solve specific workflows, they improve with usage data, and they create switching costs through accumulated context.
Microsoft's 365 Copilot has generated roughly $2B in annualised revenue, well below the $12B that Wall Street analysts projected. The gap highlights a persistent challenge: enterprises are willing to pilot AI tools but slow to roll them out across entire organisations. Usage data shows that only 15-20% of licensed Copilot users are active on a weekly basis.
We are seeing inference costs drop 10x every 18 months. That changes the economics of every AI deployment. The question is no longer whether you can afford to use AI, but whether you can afford not to. The companies building on this cost curve are pulling away from those still doing manual processes.
The most important chart in AI right now: inference cost per million tokens has dropped 95% in 18 months. This is not incremental improvement. This is a phase change. Every business case that did not work at $10/M tokens works at $0.50/M tokens. Every automation that was too expensive is now cheap. 1/12
After two years of skepticism about AI application companies, we are seeing clear evidence that the application layer is working. The best AI applications combine three things: a s...
Microsoft's 365 Copilot has generated roughly $2B in annualised revenue, well below the $12B that Wall Street analysts projected. The gap highlights a persistent challenge: enterpr...
Interesting pattern in enterprise AI procurement: companies are consolidating from 8-12 AI vendors down to 3-4 strategic partners. The experimentation phase is over. Procurement wa...
The most important chart in AI right now: inference cost per million tokens has dropped 95% in 18 months. This is not incremental improvement. This is a phase change. Every busines...
AI agents are the new aggregators. Just as Google aggregated web content and Facebook aggregated social connections, AI agents will aggregate services and workflows. The company th...
Every industry will be transformed by AI. But the transformation will not look like what people expect. It will not be AI replacing humans. It will be AI enabling humans to do thin...
Public cloud companies with meaningful AI revenue are now trading at 18x forward revenue vs 10x for those without. The gap has widened from 1.2x a year ago to 1.8x today. AI is no ...
Act One of generative AI was about model capabilities and the infrastructure to serve them. Act Two is about applications that generate sustainable revenue. We are now seeing the f...
We are seeing inference costs drop 10x every 18 months. That changes the economics of every AI deployment. The question is no longer whether you can afford to use AI, but whether y...
What NVIDIA has done is create a full-stack platform where the GPU is just the foundation. CUDA, TensorRT, Triton Inference Server -- these create switching costs that are the envy...