3D7D
 Price:

AI & TECH

Intercom and Cursor ditch frontier AI APIs for cheaper, smarter models

Wednesday, April 29, 2026 · from 2 podcasts
  • Intercom’s dedicated AI model now beats GPT-4 and Opus 4.5 in customer service at lower cost.
  • Cursor and Decagon are moving most traffic to in-house, fine-tuned open-source models.
  • The advantage shifted from raw compute to proprietary 'last-mile' user interaction data.
  • This ends the era where enterprise AI meant paying a premium for general-purpose APIs.

The ‘AI subsidy’ era is over. Six weeks after a prior report predicted AI’s disruption of corporate moats, the trend is playing out in real time.

Intercom CEO Eoin Mac Caba just claimed the top spot in customer service AI. His new Apex model reportedly beats GPT-4 and Opus 4.5 on resolution rates while slashing costs. Cursor followed a similar path in coding, with its Composer 2 model built on open-source weights now rivaling Anthropic's Opus 4.6.

“We're seeing a shift from the generalist era to the vertical era.”

- Nathaniel Whittemore, The AI Daily Brief

This is a classic disruption pattern. Frontier labs like OpenAI over-serve the market with expensive, general intelligence, while leaner companies pick off profitable verticals. According to Nathaniel Whittemore on The AI Daily Brief, companies like Pinterest and Notion already find it faster and cheaper to train and run open models themselves.

The source of power has shifted. Andrej Karpathy predicted this AI model speciation, where smaller, task-specific models thrive. The key is not raw compute or internet-scale data, but high-quality ‘last-mile’ interaction data that frontier labs cannot access. Intercom’s Apex uses proprietary customer service logs, and Cursor’s Composer 2 leverages developer feedback loops.

This redefines the ‘bitter lesson’ - Rich Sutton’s axiom that computation beats human design. He now argues learning from experience is the next phase. Vertical winners are scaling learning on human feedback, not just scaling parameters on static text.

The economic logic is clear. Decagon co-founder Ashwin Srinivasan reports 80% of their traffic runs on internal models. Paying the API tax to resell another company’s compute is now a luxury, not a necessity. The frontier labs face a choice: build cheaper specialized models themselves or watch their most lucrative customers walk away.

Source Intelligence

- Deep dive into what was said in the episodes

The AI Subsidy Era is OverApr 28

  • Intercom's new dedicated customer service model Finn Apex achieves the highest performance, speed, and cost metrics, beating GPT-4 and Opus 4.5, according to CEO Eoin Mac Caba.
  • The 'bitter lesson' from Rich Sutton argues that general methods leveraging computation beat human-designed domain-specific approaches every time. This pattern held with Bloomberg's specialized finance model being surpassed by generalist LLMs.
  • A new hypothesis challenges the bitter lesson, suggesting high-quality 'last-mile' user interaction data can make vertical models outperform frontier models through targeted post-training, not full pretraining.
  • Eoin Mac Caba claims Intercom's Apex model has a 2.8% higher resolution rate and a 65% reduction in hallucinations compared to other models, enabled by proprietary customer service interaction data.
  • Industry observers like Ben Avogi and Clem Delangue argue vertical SaaS companies with labeled interaction data have untapped fine-tuning assets, predicting a shift from API reliance to in-house open models.
  • Andrej Karpathy predicts AI model speciation, analogous to animal kingdom diversity, where smaller, task-specific models with a cognitive core will thrive over a single general oracle.
  • Richard Sutton, on the Dwarkesh podcast, framed learning from experience as the next phase of the bitter lesson, which aligns with the post-training from real interaction data seen with Apex and Composer 2.
  • Nathaniel Whittemore argues frontier AI labs face classic disruption and may need to build cheaper specialized models themselves, potentially through data partnerships or acquiring companies with proprietary evals.
Also from this episode: (1)

AI & Tech (1)

  • Cursor's Composer 2 model, based on an open-source Kimmy 2.5 with extra reinforcement learning, reportedly beats Opus 4.6 on coding benchmarks while being cheaper, showing post-training's potential.

Coffee Date with LNbits | FREEDOM TECH FRIDAY 38Apr 27

  • LNbits is a self-hosted, MIT-licensed software layer that adds an API and UI to a Lightning funding source, enabling extended functionality beyond a basic wallet.
  • LNbits offers a commercial SaaS product at mylnbits.com, allowing users to spin up hosted instances, and extensions can be monetized through one-time payments, subscriptions, or revenue-sharing models like a point-of-sale kickback.
Also from this episode: (8)

Protocol (3)

  • Black Coffee describes LNbits as an open platform akin to WordPress for a Lightning node, with a core user base of merchants, developers, and individuals running it for personal or family use.
  • The LNbits Box is a dedicated hardware product running NixOS that offers plug-and-play setup with integrated funding sources like Spark, Phoenix, or Arc, a web-based admin panel, and encrypted backup systems.
  • Black Coffee argues the self-custody narrative in Bitcoin is often a 'no true Scotsman' fallacy, advocating for pragmatic trade-offs like using Spark or Phoenix which offer usability despite not being perfectly trust-minimized.

Lightning (5)

  • LNbits supports over twenty funding sources including Core Lightning, LND, Phoenix, Spark, and the upcoming Arc, and users can switch between them at any time within the interface.
  • The platform's TunnelMeOut extension provides a paid reverse proxy service starting at 100 sats per day, granting Clearnet access to instances running on home servers without requiring manual VPS setup.
  • Major LNbits extensions include a point-of-sale system with inventory management, a webshop plugin, Nostr integrations for zaps and Lightning addresses, and channel management tools for CLN and LND backends.
  • Black Coffee says LNbits is designed for LLM-assisted 'vibe coding', providing comprehensive API documentation and an OpenAPI spec to enable rapid development of custom Lightning applications.
  • Black Coffee identifies awareness as LNbits's primary growth challenge, noting the project's broad functionality makes it harder to market than single-purpose tools.