Comment by plasticeagle

Comment by plasticeagle 18 hours ago

49 replies

ChatGPT and Gemini literally only know the answer because they read StackOverflow. Stack Overflow only exists because they have visitors.

What do you think will happen when everyone is using the AI tools to answer their questions? We'll be back in the world of Encyclopedias, in which central authorities spent large amounts of money manually collecting information and publishing it. And then they spent a good amount of time finding ways to sell that information to us, which was only fair because they spent all that time collating it. The internet pretty much destroyed that business model, and in some sense the AI "revolution" is trying to bring it back.

Also, he's specifically talking about having a coding tool write the code for you, he's not talking about using an AI tool to answer a question, so that you can go ahead and write the code yourself. These are different things, and he is treating them differently.

socalgal2 17 hours ago

> ChatGPT and Gemini literally only know the answer because they read StackOverflow. Stack Overflow only exists because they have visitors.

I know this isn't true because I work on an API that has no answers on stackoverflow (too new), nor does it have answers anywhere else. Yet, the AI seems to able to accurately answer many questions about it. To be honest I've been somewhat shocked at this.

  • gwhr 12 hours ago

    What kind of API is it? Curious if it's a common problem that the AI was able to solve?

  • bbarnett 16 hours ago

    It is absolutely true, and AI cannot think, reason, comprehend anything it has not seen before. If you're getting answers, it has seen it elsewhere, or it is literally dumb, statistical luck.

    That doesn't mean it knows the answer. That means it guessed or hallucinated correctly. Guessing isn't knowing.

    edit: people seem to be missing my point, so let me rephrase. Of course AIs don't think, but that wasn't what I was getting at. There is a vast difference between knowing something, and guessing.

    Guessing, even in humans, is just the human mind statistically and automatically weighing probabilities and suggesting what may be the answer.

    This is akin to what a model might do, without any real information. Yet in both cases, there's zero validation that anything is even remotely correct. It's 100% conjecture.

    It therefore doesn't know the answer, it guessed it.

    When it comes to being correct about a language or API that there's zero info on, it's just pure happenstance that it got it correct. It's important to know the differences, and not say it "knows" the answer. It doesn't. It guessed.

    One of the most massive issues with LLMs is we don't get a probability response back. You ask a human "Do you know how this works", and an honest and helpful human might say "No" or "No, but you should try this. It might work".

    That's helpful.

    Conversely a human pretending it knows and speaking with deep authority when it doesn't is a liar.

    LLMs need more of this type of response, which indicates certainty or not. They're useless without this. But of course, an LLM indicating a lack of certainty, means that customers might use it less, or not trust it as much, so... profits first! Speak with certainty on all things!

    • demosthanos 12 hours ago

      This is wrong. I write toy languages and frameworks for fun. These are APIs that simply don't exist outside of my code base, and LLMs are consistently able to:

      * Read the signatures of the functions.

      * Use the code correctly.

      * Answer questions about the behavior of the underlying API by consulting the code.

      Of course they're just guessing if they go beyond what's in their context window, but don't underestimate context window!

      • bbarnett 12 hours ago

        So, you're saying you provided examples of the code and APIs and more, in the context window, and it succeeds? That sounds very much unlike the post I responded to, which claimed "no knowledge". You're also seemingly missing this:

        "If you're getting answers, it has seen it elsewhere"

        The context window is 'elsewhere'.

    • lechatonnoir 16 hours ago

      This is such a pointless, tired take.

      You want to say this guy's experience isn't reproducible? That's one thing, but that's probably not the case unless you're assuming they're pretty stupid themselves.

      You want to say that it Is reproducible, but that "that doesn't mean AI can think"? Okay, but that's not what the thread was about.

    • hombre_fatal 15 hours ago

      This doesn't seem like a useful nor accurate way of describing LLMs.

      When I built my own programming language and used it to build a unique toy reactivity system and then asked the LLM "what can I improve in this file", you're essentially saying it "only" could help me because it learned how it could improve arbitrary code before in other languages and then it generalized those patterns to help me with novel code and my novel reactivity system.

      "It just saw that before on Stack Overflow" is a bad trivialization of that.

      It saw what on Stack Overflow? Concrete code examples that it generalized into abstract concepts it could apply to novel applications? Because that's the whole damn point.

      • skydhash 13 hours ago

        Programming languages, by their nature of being formal notation, only have a few patterns to follow, all of them listed in the grammar of that language. And then there’s only so much libraries out there. I believe there’s more unique comments and other code explanations out there than unique code patterns. Take something like MDN where there’s a full page of text for every JavaScript, html, css symbol.

    • Workaccount2 10 hours ago

      >It is absolutely true, and AI cannot think, reason, comprehend anything it has not seen before. If you're getting answers, it has seen it elsewhere, or it is literally dumb, statistical luck.

      How would you reconcile this with the fact that SOTA models are only a few TB in size? Trained on exabytes of data, yet only a few TB in the end.

      Correct answers couldn't be dumb luck either, because otherwise the models would pretty much only hallucinate (the space of wrong answers is many orders of magnitude larger than the space of correct answers), similar to the early proto GPT models.

      • efavdb 10 hours ago

        Could it be that there is a lot of redundancy in the training data?

      • daveguy 9 hours ago

        > How would you reconcile this with the fact that SOTA models are only a few TB in size? Trained on exabytes of data, yet only a few TB in the end.

        This is false. You are off by ~4 orders of magnitude by claiming these models are trained on exabytes of data. It is closer to 500TB of more curated data at most. Contrary to popular belief LLMs are not trained on "all of the data on the internet". I responded to another one of your posts that makes this false claim here:

        https://news.ycombinator.com/item?id=44283713

    • PeterStuer 16 hours ago

      What would convince you otherwise? The reason I ask is because you sound like you have made up your mind phylosophically, not based on practical experience.

    • rsanheim 16 hours ago

      It's just Pattern matching. Most APIs, and hell, most code is not unique or special. Its all been done a thousands of times before. Thats why an LLM can be helpful on some tool you've written just for yourself and never released anywhere.

      As to 'knows the answer', I'm don't even know what that means with these tools. All I know is if it is helpful or not.

      • danielbln 13 hours ago

        Also, most problems are decomposable into simpler, certainly not novel parts. That intractable unicorn problem I hear so much about is probably composed of very pedestrian sub-problems.

      • CamperBob2 9 hours ago

        'Pattern matching' isn't just all you need, it's all there is.

    • jumploops 15 hours ago

      > It is absolutely true, and AI cannot think, reason, comprehend anything it has not seen before.

      The amazing thing about LLMs is that we still don’t know how (or why) they work!

      Yes, they’re magic mirrors that regurgitate the corpus of human knowledge.

      But as it turns out, most human knowledge is already regurgitation (see: the patent system).

      Novelty is rare, and LLMs have an incredible ability to pattern match and see issues in “novel” code, because they’ve seen those same patterns elsewhere.

      Do they hallucinate? Absolutely.

      Does that mean they’re useless? Or does that mean some bespoke code doesn’t provide the most obvious interface?

      Having dealt with humans, the confidence problem isn’t unique to LLMs…

      • skydhash 13 hours ago

        > The amazing thing about LLMs is that we still don’t know how (or why) they work!

        You may want to take a course in machine learning and read a few papers.

      • rainonmoon 12 hours ago

        > the corpus of human knowledge.

        Goodness this is a dim view on the breadth of human knowledge.

    • [removed] 16 hours ago
      [deleted]
    • gejose 10 hours ago

      I'm sorry but this is a gross oversimplification. You can also apply this to the human brain.

      "<the human brain> cannot think, reason, comprehend anything it has not seen before. If you're getting answers, it has seen it elsewhere, or it is literally dumb, statistical luck."

semiquaver 9 hours ago

> ChatGPT and Gemini literally only know the answer because they read StackOverflow

Obviously this isn’t true. You can easily verify this by inventing and documenting an API and feeding that description to an LLM and asking it how to use it. This works well. LLMs are quite good at reading technical documentation and synthesizing contextual answers from it.

erikerikson 10 hours ago

I broadly agree that cutting new knowledge will need to continue being done and that overuse of LLMs could undermine that, yet... When was the last time you paid to read an APIs' docs? It costs money for companies to make those too.

Taylor_OD 10 hours ago

> ChatGPT and Gemini literally only know the answer because they read StackOverflow. Stack Overflow only exists because they have visitors.

I mean... They also can read actual documentation. If I'm working on any api work or a language I'm not familiar with, I ask the LLM to include the source they got their answer from and use official documentation when possible.

That lowers the hallucination rate significantly and also lets me ensure said function or code actually does what the llm reports it does.

In theory, all stackoverflow answers are just regurgitated documentation, no?

  • sothatsit 9 hours ago

    > I mean... They also can read actual documentation.

    This 100%. I use o3 as my primary search engine now. It is brilliant at finding relevant sources, summarising what is relevant from them, and then also providing the links to those sources so I can go read them myself. The release of o3 was a turning point for me where it felt like these models could finally go and fetch information for themselves. 4o with web search always felt inadequate, but o3 does a very good job.

    > In theory, all stackoverflow answers are just regurgitated documentation, no?

    This is unfair to StackOverflow. There is a lot of debugging and problem solving that has happened on that platform of undocumented bugs or behaviour.

olmo23 16 hours ago

Where does the knowledge come from? People can only post to SO if they've read the code or the documentation. I don't see why LLMs couldn't do that.

  • nobunaga 16 hours ago

    ITT: People who think LLMs are AGI and can produce output that the LLM has come up with out of thin air or by doing research. Go speak with someone who is actually an expert in this field how LLMs work and why the training data is so important. Im amazed that people in the CS industry seem to talk like they know everything about a tech after using it but never even writing a line of code for an LLM. Our indsutry is doomed with people like this.

    • usef- 15 hours ago

      This isn't about being AGI or not, and it's not "out of thin air".

      Modern implementations of LLMs can "do research" by performing searches (whose results are fed into the context), or in many code editors/plugins, the editor will index the project codebase/docs and feed relevant parts into the context.

      My guess is they either were using the LLM from a code editor, or one of the many LLMs that do web searches automatically (ie. all of the popular ones).

      They are answering non-stackoverflow questions every day, already.

      • nobunaga 11 hours ago

        Yeah, doing web searches could be called research but thats not what we are talking bout. Read the parent of the parent. Its about being able to answer questions thats not in its training data. People are talking about LLMs making scientific discoveries that humans haven't. A ridiculous take. Its not possible and with the current state of tech never will be. I know what LLMs are trained on. Thats not the topic of conversation.

    • planb 14 hours ago

      I think the time has come to not mean LLMs when talking about AI. An agent with web access can do so much more and hallucinates way less than "just" the model. We should start seeing the model as a building block of an AI system.

    • raincole 12 hours ago

      > LLM has come up with out of thin air

      People don't think that. Especially not the commentor you replied to. You're human-hallucinating.

      People think LLM are trained on raw documents and code besides StackOverflow. Which is very likely true.

      • nobunaga 11 hours ago

        Read the parent of the parent. Its about being able to answer questions thats not in its training data. People are talking about LLMs making scientific discoveries that humans havent. A ridiculous take. Its not possible and with the current state of tech never will be. I know what LLMs are trained on. Thats not the topic of conversation.

reaperducer 7 hours ago

We'll be back in the world of Encyclopedias

On a related note, I recently learned that you can still subscribe to the Encyclopedia Britannica. It's $9/month, or $75/year.

Considering the declining state of Wikipedia, and the untrustworthiness of A.I., I'm considering it.

CamperBob2 9 hours ago

We'll start writing documentation for primary consumption by LLMs rather than human readers. The need for sites like SO will not vanish overnight but it will diminish drastically.

kypro 16 hours ago

The idea that LLMs can only spew out text they've been trained on is a fundamental miss-understanding of how modern backprop training algorithms work. A lot of work goes into refining training algorithms to preventing overfitting of the training data.

Generalisation is something that neural nets are pretty damn good at, and given the complexity of modern LLMs the idea that they cannot generalise the fairly basic logical rules and patterns found in code such that they're able provide answers to inputs unseen in the training data is quite an extreme position.

  • fpoling 11 hours ago

    Yet the models do not (yet) reason. Try to ask them to solve a programming puzzle or exercise from an old paper book that was not scanned. They will produce total garbage.

    Models work across programming languages because it turned out programming languages and API are much more similar than one could have expected.