Comment by infecto

Comment by infecto 3 days ago

13 replies

If you ask an LLM what color is the sky it might say purple but if you give it a paragraph describing the atmosphere and then ask the same question it will almost always answer correctly. I don't think hallucinations are as big of a problem as people make them out to be.

misnome 3 days ago

So, it only works if you already know enough about the problem to not need to ask the LLM, check.

  • infecto 3 days ago

    Are you just writing negative posts without even seeing the product? The system queries the internet, aggregates that information and writes information based on your query.

    • misnome 2 days ago

      ChatGPT, please explain threaded discussions and context of statements as if you were talking to a five year old.

      • infecto 2 days ago

        Ahh so you are a child who has no intellectual capability past writing negative attack statements. Got it.

  • keiferski 3 days ago

    No, if the data you’re querying contains the information you need, then it is mostly fine to ask for that data in a format amendable to your needs.

    • o11c 3 days ago

      The problem with LLMs is not a data problem. LLMs are stupid even on data they just generated.

      One recent catastrophic failure I found: Ask an LLM to generate 10 pieces of data. Then in a second input, ask it to select (say) only numbers 1, 3, and 5 from the list. The LLM will probably return results numbered 1, 3, and 5, but chances are at least one of them will actually copy the data from a different number.

      • wsve 2 days ago

        I'm absolutely not bullish on LLMs, but I think this is kinda judging a fish on its ability to climb a tree.

        LLMs are looking at typical constructions of text, not an understanding of what it means. If you ask it what color the sky is, it'll find what text usually follows a sentence like that, and tries to construct a response from it.

        If you ask it the answer to a math question, the only way it could reliably figure it out is if it has in its database an exact copy of that math question. Asking it to choose things from a list is kinda like that, but one could imagine that the designers would try to supplement that manually with a different technique from pure LLM.

      • smcin 3 days ago

        Any ideas why that misnumbering happens? It sounds a very basic thing to get wrong. And as a fallback, could be brute-force kludged with an extra pass which appends the output list to the prompt.

        • o11c 3 days ago

          It's an LLM, we cannot expect any real idea.

          Unless of course we rephrase it as "when I roll 2d6, why do I sometimes get snake eyes?"

chx 3 days ago

There are no hallucinations. It's just the normal bullshit people hang a more palatable name on. There is nothing else.

https://hachyderm.io/@inthehands/112006855076082650

> You might be surprised to learn that I actually think LLMs have the potential to be not only fun but genuinely useful. “Show me some bullshit that would be typical in this context” can be a genuinely helpful question to have answered, in code and in natural language — for brainstorming, for seeing common conventions in an unfamiliar context, for having something crappy to react to.

> Alas, that does not remotely resemble how people are pitching this technology.

infecto 3 days ago

Why does this get downvoted so heavily? It’s my experience running LLM in production. At scale hallucinations are not a huge problem when you have reference material.