Comment by dotancohen

Comment by dotancohen 10 months ago

19 replies

Yes, but the API at the end is providing the core functionality. Simply swapping out one LLM model for another - let alone by a different company altogether - will completely change the effectiveness and usefulness of the application.

Tepix 10 months ago

Well, as we see with AI applications like "Leo AI" and "Continue", using a locally run LLM can be fantastic replacements for proprietary offerings.

  • dartos 10 months ago

    FWIW I’ve found local models to be essentially useless for coding tasks.

    • Tepix 10 months ago

      Really? Maybe your models are too small?

      • spmurrayzzz 10 months ago

        The premier open weight models don't even comparatively perform well on the public benchmarks compared to frontier models. And that's assuming at least some degree of benchmark contamination for the open weight models.

        While I don't think they're completely useless (though its close), calling them fantastic replacements feels like an egregious overstatement of their value.

        EDIT: Also wanted to note that I think this becomes as much an expectations-setting exercise as it is evaluation on raw programming performance. Some people are incredibly impressed by the ability to assist in building simple web apps, others not so much. Experience will vary across that continuum.

        • dartos 10 months ago

          Yeah, in my comparing deepseek coder 2 lite (the best coding model I can find that’ll run on my 4090) to Claud sonnet under aider…

          Deep seek lite was essentially useless. Too slow and too low quality edits.

          I’ve been programming for about 17 years, so the things I want aider to do are a little more specific than building simple web apps. Larger models are just better at it.

          I can run the full deepseek coder model on some cloud and probably get very acceptable results, but then it’s no longer local.

      • websap 10 months ago

        Woah woah! Those are fighting words. /s

dartos 10 months ago

One would hope, that since the problem these models are trying to solve is language modeling, they would eventually converge around similar capabilities

JCharante 10 months ago

everyone stands on the shoulders of giants.

  • sham1 10 months ago

    Things standing on the shoulders of proprietary giants shouldn't claim to be free software/open source.

    • t-writescode 10 months ago

      Their interfacing software __is__ open source; and, they're asking for your OpenAI api key to operate. I would expect / desire open source code if I were to use that, so I could be sure my api key was only being used for my work, so it's only my work that I'm paying for and it's not been stolen in some way.

  • noduerme 10 months ago

    My older brother who got me into coding learned to code in Assembly. He doesn't really consider most of my work writing in high level languages to be "coding". So maybe there's something here. But if I had to get into the underlying structure, I could. I do wonder whether the same can be said for people who just kludge together a bunch of APIs that produce magical result sets.

    • dotancohen 10 months ago

        > But if I had to get into the underlying structure, I could.
      
      How do you propose to get into the underlying structure of the OpenAPI API? Breach their network and steal their code and models? I don't understand what you're arguing.
      • latexr 10 months ago

        > How do you propose to get into the underlying structure of the OpenAPI API?

        The fact that you can’t is the point of the comment. You could get into the underlying structure of other things, like the C interpreter of a scripting language.

      • K0balt 10 months ago

        I think the relevant analogy here would be to run a local model. There are several tools to easily run local models for a local API. I run a 70b finetune with some tool use locally on our farm, and it is accessible to all users as a local openAI alternative. For most applications it is adequate and data stays on the campus area network.

      • seadan83 10 months ago

        I think the argument is that stitching things together at a high level is not really coding. A bit of a no true scotsmen perspective. The example is that anything more abstract than assembly is not even true coding, let alone creating a wrapper layer around an LLM