Comment by gonzalohm

Comment by gonzalohm 3 days ago

18 replies

Probably a lot of people here disagree with this feeling. But my take is that if setting up all the AI infrastructure and onboarding to my code is going to take this amount of effort, then I might as well code the damn thing myself which is what I'm getting paid to (and enjoy doing anyway)

fragmede 3 days ago

Whether it's setting up AI infrastructure or configuring Emacs/vim/VSCode, the important distinction to make is if the cost has to be paid continually, or if it's a one time/intermittent cost. If I had to configure my shell/git aliases every time I booted my computer, I wouldn't use them, but seeing as how they're saved in config files, they're pretty heavily customized by this point.

Don't use AI if you don't want to, but "it takes too much effort to set up" is an excuse printf debuggers use to avoid setting up a debugger. Which is a whole other debate though.

  • bird0861 2 days ago

    I fully agree with this POV but for one detail; there is a problem with sunsetting frontier models. As we begin to adopt these tools and build workflows with them, they become pieces of our toolkit. We depend on them. We take them for granted even. And then the model either changes (new checkpoints, maybe alignment gets fiddled with) and all of the sudden prompts no longer yield the same results we expected from them after working on them for quite some time. I think the term for this is "prompt instability". I felt this with Gemini 3 (and some people had less pronounced but similar experience with Sonnet releases after 3.7) which for certain tasks that 2.5Pro excelled at..it's just unusable now. I was already a local model advocate before this but now I'm a local model zealot. I've stopped using Gemini 3 over this. Last night I used Qwen3 VL on my 4090 and although it was not perfect (sycophancy, overuse of certain cliches...nothing I can't get rid of later with some custom promptsets and a few hours in Heretic) it did a decent enough job of helping me work through my blindspots in the UI/UX for a project that I got what I needed.

    If we have to perform tuning on our prompts ("skills", agents.md/claude.md, all of the stuff a coding assistant packs context with) every model release then I see new model releases becoming a liability more than a boon.

Havoc 3 days ago

A lot of the style stuff you can write once and reuse. I started splitting mine into overall and project specific files for this reason

Universal has stuff I always want (use uv instead of pip etc) while the other describes what tech choice for this project

vanviegen 3 days ago

Perhaps. But keep in mind that the setup work is typically mostly delegated to LLMs as well.

kissgyorgy 3 days ago

I strongly disagree with the author not using /init. It takes a minute to run and Claude provides surprisingly good results.

  • 0xblacklight 3 days ago

    If you find it works for you, then that’s great! This post is mostly from our learnings from getting it to solve hard problems in complex brownfield codebases where auto generation is almost never sufficient.

  • alwillis 3 days ago

    /init has evolved since the early day; it's more concise than it used to be.

nvarsj 3 days ago

It really doesn't take that much effort. Like any tool, people can over-optimise on the setup rather than just use it.

nichochar 3 days ago

The effort described in the article is maybe a couple hours of work.

I understand the "enjoy doing anyway" part and it resonates, but not using AI is simply less productive.

  • globular-toast 2 days ago

    It's a couple of hours right now, then another couple of hours "correcting" the AI when it still goes wrong, another couple of hours tweaking the file again, another couple of hours to update when the model changes, another couple of hours when someone writes a new blog post with another method etc.

    There's a huge difference between investing time into a deterministic tool like a text editor or programming language and a moving target like "AI".

    The difference between programming in Notepad in a language you don't know and using "AI" will be huge. But the difference between being fluent in a language and having a powerful editor/IDE? Minimal at best. I actually think productivity is worse because it tricks you into wasting time via the "just one more roll" (ie. gambling) mentality. Not to mention you're not building that fluency or toolkit for yourself, making you barely more valuable than the "AI" itself.

    • fragmede 2 days ago

      You say that as if tech hasn't always been a moving target anyway. The skills I spent months learning a specific language and IDE became obsolete with the next job and the next paradigm shift. That's been one of the few consistent themes throughout my career. Hours here and there, spread across months and years, just learning whatever was new. Sometimes, like with Linux, it really paid off. Other times, like PHP, it did, and then fizzled out.

      --

      The other thing is, this need for determinism bewilders me. I mean, I get where it comes from, we want nice, predictable reliable machines. But how deterministic does it need to be? If today, it decides to generate code and the variable is called fileName, and tomorrow it's filePath, as long as it's passing tests, what do I care that it's not totally deterministic and the names of the variables it generates are different? as long as it's consistent with existing code, and it passes tests, whats the importance of it being deterministic to a computer science level of rigor? It reminds me about the travelling salesman problem, or the knapsack problem. Both NP hard, but users don't care about that. They just want the computer to tell them something good enough for them to go on about their day. So if a customer comes up to you and offers you a pile of money to solve either one of those problems, do I laugh in their face, knowing damn well I won't be the one to prove that NP = P, or do I explain to them the situation, and build them software that will do the best it can, with however much compute resources they're willing to pay for?

  • TheRoque 3 days ago

    > but not using AI is simply less productive

    Some studies shows the opposite for experienced devs. And it also shows that developers are delusional about said productivity gains: https://metr.org/blog/2025-07-10-early-2025-ai-experienced-o...

    If you have a counter-study (for experienced devs, not juniors), I'd be curious to see. My experience also has been that using AI as part of your main way to produce code, is not faster when you factor in everything.

    • ares623 2 days ago

      Curious why there hasn't been a rebuttal study to that one yet (or if there is I haven't seen it come up). There must be near infinite funding available to debunk that study right?

    • bird0861 2 days ago

      That study is garbo and I suspect you didn't even read the abstract. Am I right?

      • gravypod 2 days ago

        I've heard this mentioned a few times. Here is a summarized version of the abstract:

            > ... We conduct a randomized controlled trial (RCT)
            > ... AI tools ... affect the productivity of experienced
            > open-source developers. 16 developers with moderate AI
            > experience complete 246 tasks in mature projects on which they
            > have an average of 5 years of prior experience. Each task is
            > randomly assigned to allow or disallow usage of early-2025 AI
            > tools. ... developers primarily use Cursor Pro ... and
            > Claude 3.5/3.7 Sonnet. Before starting tasks, developers forecast that allowing
            > AI will reduce completion time by 24%. After completing the
            > study, developers estimate that allowing AI reduced completion time by 20%.
            > Surprisingly, we find that allowing AI actually increases
            > completion time by 19%—AI tooling slowed developers down. This
            > slowdown also contradicts predictions from experts in economics
            > (39% shorter) and ML (38% shorter). To understand this result,
            > we collect and evaluate evidence for 21 properties of our setting
            > that a priori could contribute to the observed slowdown effect—for
            > example, the size and quality standards of projects, or prior
            > developer experience with AI tooling. Although the influence of
            > experimental artifacts cannot be entirely ruled out, the robustness
            > of the slowdown effect across our analyses suggests it is unlikely
            > to primarily be a function of our experimental design.
        
        So what we can gather:

        1. 16 people were randomly given tasks to do

        2. They knew the codebase they worked on pretty well

        3. They said AI would help them work 24% faster (before starting tasks)

        4. They said AI made them ~20% faster (after completion of tasks)

        5. ML Experts claim that they think programmers will be ~38% faster

        6. Economists say ~39% faster.

        7. We measured that people were actually 19% slower

        This seems to be done on Cursor, with big models, on codebases people know. There are definitely problems with industry-wide statements like this but I feel like the biggest area AI tools help me is if I'm working on something I know nothing about. For example: I am really bad at web development so CSS / HTML is easier to edit through prompts. I don't have trouble believing that I would be slower trying to make an edit to code that I already know how to make.

        Maybe they would see the speedups by allowing the engineer to select when to use the AI assistance and when not to.

  • svachalek 2 days ago

    Minutes really, despite what the article says you can get 90% of the way there by telling Claude how you want the project documentation structured and just let it do it. Up to you if you really want to tune the last 10% manually, I don't. I have been using basically the same system and when I tell Claude to update docs it doesn't revert to one big Claude.md, it maintains it in a structure like this.