Comment by deadbabe

Comment by deadbabe 5 days ago

7 replies

It took three days because... agents suck.

But yes, with enough prodding they will eventually build you something that's been built before. Don't see why that's particularly impressive. It's in the training data.

simonw 5 days ago

Not a useful mental model.

  • deadbabe 5 days ago

    It is useful. If you can whip up something complex fairly quickly with an AI agent, it’s likely because it’s already been done before.

    But if even the AI agent seems to struggle, you may be doing something unprecedented.

    • simonw 5 days ago

      Except if you spend quality time with coding agents you realize that's not actually true.

      They're equally useful for novel tasks because they don't work by copying large scale patterns from their training data - the recent models can break down virtually any programming task to a bunch of functions and components and cobble together working code.

      If you can clearly define the task, they can work towards a solution with you.

      The main benefit of concepts already in the training data is that it lets you slack off on clearly defining the task. At that point it's not the model "cheating", it's you.

      • deadbabe 4 days ago

        Good long lived software is not a bunch of functions and components cobbled together.

        You need to see the big picture and visions of the future state in order to ensure what is being built will be able to grow and breathe into that. This requires an engineer. An agent doesn’t think much about the future, they think about right now.

        This browser toy built by the agent, it has NO future. Once it has written the code, the story is over.

      • aix1 5 days ago

        Simon, do you happen to have some concrete examples of a model doing a great job at a clearly novel, clearly non-trivial coding task?

        I'd find it very interesting to see some compelling examples along those line.

      • keybored 5 days ago

        > Except if you spend quality time with coding agents you realize that's not actually true.

        Agent engineering seems to be (from the outside!) converging on quality lived experience. Compared to Stone Age manual coding it’s less about technical arguments and more about intuition.

        Vibes in short.

        You can’t explain sex to someone who has not had sex.

        Any interaction with tools is partly about intuition. It’s a difference of degree.