Comment by voidspark

Comment by voidspark 2 days ago

6 replies

But that's exactly what these deep neural networks have shown, countless times. LLM's generalize on new data outside of its training set. It's called "zero shot learning" where they can solve problems that are not in their training set.

AlphaGo Zero is another example. AlphaGo Zero mastered Go from scratch, beating professional players with moves it was never trained on

> Another is the fundamental inability to self update

That's an engineering decision, not a fundamental limitation. They could engineer a solution for the model to initiate its own training sequence, if they decide to enable that.

no_wizard 2 days ago

>AlphaGo Zero mastered Go from scratch, beating professional players with moves it was never trained on

Thats all well and good, but it was tuned with enough parameters to learn via reinforcement learning[0]. I think The Register went further and got better clarification about how it worked[1]

>During training, it sits on each side of the table: two instances of the same software face off against each other. A match starts with the game's black and white stones scattered on the board, placed following a random set of moves from their starting positions. The two computer players are given the list of moves that led to the positions of the stones on the grid, and then are each told to come up with multiple chains of next moves along with estimates of the probability they will win by following through each chain.

While I also find it interesting that in both of these instances, its all referenced to as machine learning, not AI, its also important to see that even though what AlphaGo Zero did was quite awesome and a step forward in using compute for more complex tasks, it was still seeded the basics of information - the rules of Go - and simply patterned matched against itself until built up enough of a statistical model to determine the best moves to make in any given situation during a game.

Which isn't the same thing as showing generalized reasoning. It could not, then, take this information and apply it to another situation.

They did show the self reinforcement techniques worked well though, and used them for Chess and Shogi to great success as I recall, but thats a validation of the technique, not that it could generalize knowledge.

>That's an engineering decision, not a fundamental limitation

So you're saying that they can't reason about independently?

[0]: https://deepmind.google/discover/blog/alphago-zero-starting-...

[1]: https://www.theregister.com/2017/10/18/deepminds_latest_alph...

  • voidspark 2 days ago

    AlphaGo Zero didn't just pattern match. It invented moves that it had never been shown before. That is generalization, even if it's domain specific. Humans don't apply Go skills to cooking either.

    Calling it machine learning and not AI is just semantics.

    For self updating I said it's an engineering choice. You keep moving the goal posts.

    • Jensson 2 days ago

      > That is generalization, even if it's domain specific

      But that is the point, it is a domain specific AI, not a general AI. You can't train a general AI that way.

      > For self updating I said it's an engineering choice. You keep moving the goal posts.

      No, it is not an engineering choice, it is an unsolved problem to make a general AI that self updates productively. Doing that for a specific well defined problem with well defined goals is easy, but not general problem solving.

      • voidspark a day ago

        You are shifting the goal posts from AI to AGI. That's outside of the scope of the argument.

        For self updating - yes it is an engineering choice. It's already engineered in some narrow cases such as AutoML