csar 3 days ago

I’m curious how you think this compares to the Research -> Plan -> Implement method and prompts from the “Advanced Context Engineering from Agents” video when it comes to actual coding performance on large codebases. I think picking up skills is useful for broadening agents abilities, but I’m not sure I’d that’s the right thing for actual development.

The packaged collection is very cool and so is the idea of automatically adding new abilities, but I’m not fully convinced that this concept of skills is that much better than having custom commands+sub-agents. I’ll have to play around with it these next few days and compare.

  • ehsanu1 3 days ago

    Using Research->Plan->Implement flow is orthogonal, though I notice parts of those do exist as skills too. But you sometimes need to do other things too, e.g. debugging in the course of implementing or specific techniquws to improve brainstorming/researching.

    Some of these skills are probably better as programmed workflows that the LLM is forced to go through to improve reliability/consistency, that's what I've found in my own agents, rather than using English to guide the LLM and trusting it to follow the prescribed set of steps needed. Some mix of LLMs (choosing skills, executing the fuzzy parts of them) and just plain code (orchestration of skills) seems like the best bet to me and what I'm pursuing.

smrtinsert 2 days ago

Curious what you think of sub agents, don't they still consume a massive amount of tokens compared to simply running in main context? I'm skeptical of any process that starts massively delegating to sub agents. I'm on Pro and don't think its worth upgrading to 200 a month just to not pollute main context.

  • redhale 2 days ago

    In my opinion, subagents (or more generally, "agents as tools" as a pattern) are an order-of-magnitude level feature. Soon every CLI agent will have them as a first-class feature (you can get them via custom scripting right now with and CLI agent, albeit less ergonomically).

    The ability to isolate context-noisy subtasks (like agentically searching through a large codebase by grepping through dozens of irrelevant files to find the one you actually need) unlocks much longer-running loops, and therefore much more complex tasks.

    And you don't need a system this complicated to take advantage of it. Literally just a simple "codebase-searcher" agent (and Claude can vibe the agent definition for you) is enough to see the benefit first-hand. Once you see it, if you're like me, you will see opportunities for subagents everywhere.

  • simonw 2 days ago

    I think they're worthwhile only as a token context management tool - to complete side quests without using up tokens in your main agent loop.

    Using them in a way that doesn't waste tokens is something I haven't fully figured out out yet!

    • gner75 18 hours ago

      I'm not sure I get this. If anything, they'll consume less tokens, because their context will possibly contain a subset of the original single agent prompt, and they only need to see a subset of the original single agent history.

      What am I missing?

      • simonw 16 hours ago

        Take a look at my example here - having a bunch of sub-agents perform a task consumed 50,000+ tokens each across 5 subtasks, because each one had to consume duplicate information. https://simonwillison.net/2025/Oct/11/sub-agents/

        • gner75 14 hours ago

          But that's down to the way Claude Code has implemented it? If I code this myself I could engineer so that the subagents don't have overlapping context with the orchestrator.

          Also, memory itself can be a tool the subagent calls to retrieve only the stuff it needs.