Comment by wrs
> LLMs of today copy a lot of human behavior
Funny, I would say they copy almost no human behavior other than writing a continuation of an existing text.
> LLMs of today copy a lot of human behavior
Funny, I would say they copy almost no human behavior other than writing a continuation of an existing text.
LLMs aren't actually able to do that though, are they? They are simply incapable of keeping track of consistent behaviors and beliefs. I recognize that for certain prompts an LLM has to do it. But as long as we're using transformers, it'll never actually work.
People just keep underestimating transformers. Big mistake. The architecture is incredibly capable.
LLMs are capable of keeping track of consistent behaviors and beliefs, and they sure try. Are they perfect at it? Certainly not. They're pretty good at it though.
Do you understand just how much copied human behavior goes into that?
An LLM has to predict entire conversations with dozens of users, where each user has his own behaviors, beliefs and more. That's the kind of thing pre-training forces it to do.