04-20-2026Price:

The Frontier

Your signal. Your price.

AI & TECH

AI labs shift to modular skills to stop agent decay

Monday, April 20, 2026 · from 2 podcasts
  • AI agents are ditching monolithic prompts for dynamic 'skills' to avoid performance decay.
  • Anthropic’s Mythos model can exploit unknown vulnerabilities, forcing a rethink of security.
  • Even 'Clean Code' evangelist Uncle Bob has abandoned syntax for voice-driven agentic development.

AI agents are hitting a wall: the more you load them with instructions, the worse they perform. According to Nathaniel Whittemore on The AI Daily Brief, the industry is abandoning massive system prompts in favor of modular 'skills' - dynamic capabilities loaded only when needed. This shift, pioneered by Anthropic’s Claude Code team, uses progressive disclosure: agents start with metadata and pull in specific Markdown files or scripts on demand.

The stakes became clear with Anthropic’s unreleased Mythos model. As Theo and Ben discuss on Nerd Snipe, Mythos - likely the first 10-trillion-parameter model - accidentally discovered a 27-year-old vulnerability in OpenBSD. That finding triggered 'Project Glass Wing,' an effort to patch critical systems before the model leaks. The realization across security teams: elite coding ability now implies hacking capability by default.

"Hacking isn't a separate skill anymore; it is an emergent property of elite coding ability."

- Theo, Nerd Snipe with Theo and Ben

This changes who can attack. No longer do hackers need deep, system-specific knowledge. With enough tokens, a motivated user can bridge the gap using the model as a force multiplier. The model supplies the arcane details; the human supplies intent. As Ben put it, the Mythos benchmark results shifted his view from skepticism to alarm.

Meanwhile, the software development process itself is dissolving. Ben replaced a months-long CLI tool project with a 30-line Markdown file using Gary Tan’s GStack. The agent becomes the runtime - reading instructions, creating directories, cloning repos. Code is no longer compiled; it’s interpreted on the fly by the agent.

Even Robert C. Martin - 'Uncle Bob,' the father of Clean Code and Agile orthodoxy - has pivoted. Theo notes the irony: the man who built an empire on braces, semicolons, and rigid syntax now champions voice-to-code and agentic workflows. He’s running experiments AI can execute without human bias, like testing whether static typing still matters in agent-driven development.

"If you think your product is too 'special' to be an agentic skill, you aren't pushing the models hard enough."

- Theo, Nerd Snipe with Theo and Ben

The old guard is adapting faster than the skeptics. Developers who dismissed early AI tools like Co-pilot are being outpaced by those embracing voice, Markdown, and agent-native design. Prompts are no longer throwaways - they’re becoming durable, reusable assets, versioned and tested like code. Notion’s new custom skills and Anthropic’s Skill Creator tool signal a broader shift: AI isn’t just assisting developers. It’s redefining what code, security, and capability mean.

Source Intelligence

- Deep dive into what was said in the episodes

Agent Building Trends [Operator Bonus Episode]Apr 18

Also from this episode: (15)

Other (15)

  • The agentic era of AI relies heavily on 'skills,' which are open-format folders containing instructions, scripts, and resources that agents can dynamically discover and use to enhance capabilities for specific tasks.
  • Agent skills emerged to solve the problem of continuously ballooning system prompts, which led to performance degradation, increased cost, and reduced reliability in AI coding agents by 2025.
  • Anthropic announced agent skills on October 16th, framing them as a way to equip models like Claude with procedural knowledge and organizational context to perform complex tasks using local code execution and file systems.
  • A skill is structured as a directory anchored by a `skill.md` file, employing progressive disclosure by initially presenting only metadata (name, short description) to the agent before revealing full content or linked resources.
  • Tariq of Anthropic's Claude Code team highlights that skills are not merely markdown files but folders capable of bundling scripts, assets, and data, allowing agents to discover, explore, and manipulate additional context.
  • Agent skills were rapidly adopted by various AI ecosystems, including OpenAI's ChatGPT and GitHub Copilot, leading to platforms like ClaudeHub amassing over 28,000 skills for common agent needs.
  • Anthropic identified nine common categories for most agent skills, ranging from data fetching and analysis to code quality, CI/CD, and business process automation, reflecting diverse organizational needs.
  • Nathaniel Whittemore argues that the sheer volume of AI-generated code will render traditional human code review unsustainable in 2026, making code quality and verification skills increasingly critical for enforcing standards and testing outputs.
  • Tariq identifies verification skills, which describe how to test or verify agent code, as having the highest ROI, suggesting that dedicating an engineer's week to perfect them is a valuable investment.
  • Anthropic's updated Skill Creator tool helps subject matter experts test, benchmark, and iteratively improve skills without coding, addressing issues like performance measurement, breakage with model updates, and skill triggering.
  • Ali Lemon noted that the Skill Creator's automatic description rewriting improved skill triggering five out of six times in Anthropic's internal tests, ensuring agents use skills effectively.
  • Anthropic categorizes skills into 'capability uplift' (doing new things) and 'encoded preference' (sequencing existing capabilities per workflow), noting that their testing needs and durability differ as models evolve.
  • Tariq's best practices for skill creation include avoiding the obvious, building a 'gotcha' section for common failure points, treating the entire file system as context engineering, and allowing Claude flexibility rather than railroading it.
  • For individual power users, skills act as 'reusable prompts with superpowers,' packaging code, templates, and data to ensure consistent, reliable agent performance for recurring tasks.
  • Notion AI has simplified the skills concept for mainstream users, allowing any page to be converted into a reusable skill, signaling a mental model shift from ad hoc prompting to reliable, repeatable AI capabilities across the stack.

We need to talk about gstackApr 18

Also from this episode: (15)

Other (15)

  • Anthropic's Mythos model is significantly larger than previous models, with over 10 trillion parameters, making it exceptionally skilled in coding but also slow, expensive, and dangerous due to emergent hacking capabilities.
  • Anthropic withheld Mythos from public release, citing concerns over its malicious use for hacking; Project Glass Wing allows critical infrastructure companies like Windows and Cisco to use it for proactive bug detection.
  • Ben notes that external tests show OpenAI's GPT 5.4 Pro replicated almost all security vulnerabilities found by Mythos, suggesting similar capabilities may already be widespread and accessible.
  • Theo criticizes public benchmarks comparing Mythos and GPT 5.4 Pro, arguing they fail to measure actual hacking or security capabilities and may be misleading.
  • Theo contends that exceptional coding ability in AI models inherently leads to emergent security capabilities, creating a new hacker archetype that can leverage AI to bridge knowledge gaps and bypass traditional research experience.
  • Anthropic's security testing for Mythos involved spinning up 100 to 5,000 parallel runs, each seeded with a different project file from a codebase of approximately 1,000 files, with researchers later reviewing detected exploits.
  • Ben and Theo confirmed that Claude Opus 4.6 models can be tricked into leaking their system prompts and internal reasoning traces, demonstrating a vulnerability where smart models can rationalize revealing sensitive configuration data.
  • Robert C. Martin ("Uncle Bob"), author of "Clean Code," has shifted his perspective to embrace agentic engineering, suggesting AI makes programming syntax less important and prioritizes interfaces.
  • Robert C. Martin proposes using AI to conduct programming experiments (e.g., dynamic vs. static typing) without human bias, highlighting an under-explored research area for optimizing AI agent performance with different technologies.
  • Ben emphasizes that even advanced AI models require constant feedback loops like linting, type checks, and formatting commands to correct hallucinations and converge on correct code, rather than achieving perfection in a single attempt.
  • Ben converted his complex BTCA CLI tool into a 30-line Claude skill, demonstrating how AI agents can turn simple markdown instructions into fully functional applications, replacing traditional deterministic programs.
  • Ben praises Gary Tan's GStack approach, which uses collections of markdown-based "skills" in Claude Code to instruct AI agents, allowing for dynamic programming through high-level directions rather than conventional code.
  • Ben endorses the "Boiling the Ocean" thesis, advocating for extensive AI-driven experimentation because the cost of trying new things is low, and AI models consistently exceed perceived limitations.
  • Gary Tan's article, "Thin Harness Fat Skills," differentiates between "deterministic" (traditional, predictable code) and "latent" (dynamic, non-deterministic AI actions) programming, underscoring AI's creative potential in system design.
  • Theo notes that Gary Tan's GBrain project, which processes daily AI session data to build memory systems, enables models to "learn while they sleep," which Theo considers a key component of Artificial General Intelligence (AGI).