Comment by ComplexSystems

Comment by ComplexSystems 3 days ago

15 replies

Are you kidding? This is an incredible result. Stuff like this is the most important stuff happening in AI right now. Automated theorem proving? It's not too far to say the entire singular point of the technology was to get us to this.

ares623 3 days ago

This is me being snarky and ignorant, but if it solved one problem and it is automated what’s stopping it from solving all the others? That’s what’s ultimately being sold by the tweet right.

  • aeve890 3 days ago

    >This is me being snarky and ignorant, but if it solved one problem and it is automated what’s stopping it from solving all the others?

    Yeah that's the crux of the matter. How do AI did it? Using already existing math. If we need new math to prove Collatz, Goldbach or Riemman, LLMs are simply SOL. That's what's missing and hype boys always avoid to mention.

    • dwohnitmok 2 days ago

      > How do AI did it? Using already existing math. If we need new math to prove Collatz, Goldbach or Riemman, LLMs are simply SOL.

      An unproved theorem now proved is by definition new math. Will LLMs get you to Collatz, Goldbach, or Riemann? Unclear.

      But it's not like there's some magical, entirely unrelated to existing math, "new math" that was required to solve all the big conjectures of the past. They proceeded, as always, by proving new theorems one by one.

      • yorwba 2 days ago

        Yes, "new math" is neither magical nor unrelated to existing math, but that doesn't mean any new theorem or proof is automatically "new math." I think the term is usually reserved for the definition of a new kind of mathematical object, about which you prove theorems relating it to existing math, which then allows you to construct qualitatively new proofs by transforming statements into the language of your new kind of object and back.

        I think eventually LLMs will also be used as part of systems that come up with new, broadly useful definitions, but we're not there yet.

      • aeve890 2 days ago

        >An unproved theorem now proved is by definition new math.

        No. By _new math_ I mean new mathematical constructs and theories like (to mention the "newest" ones) category theory, information theory, homotopy type theory, etc. Something like Cantor inventing set theory, or Shannon with information theory, or Euler with graph theory.

        AFAIK no new field of mathematics as been created _by_ AI. Feel free to correct me.

  • ComplexSystems 3 days ago

    Ultimately the main thing that will stop it from solving literally "all the others" are things like the impossibility of solving the halting problem, considerations like P ≠ NP, etc. But as we have just seen, despite these impossibility theorems, AI systems are still able to make substantive progress on solving important open real-world problems.

    • thaumasiotes 2 days ago

      > But as we have just seen, despite these impossibility theorems, AI systems are still able to make substantive progress on solving important open real-world problems.

      You seem to be smuggling in the assumption that this problem was "important".

      • anon291 a day ago

        All problems are important. You never know what might be useful.

BanditDefender 3 days ago

I think it is way too far to say that!

We've had automated theorem proving since the 60s. What we need is automated theorem discovery. Erdős discovered these theorems even if he wasn't really able to prove them. Euler and Gauss discovered a ton of stuff they couldn't prove. It is weird that nobody considers this to be intelligence. Instead intelligence is a little game AI plays with Lean.

AI researchers keep trying to reduce intelligence into something tiny and approachable, like automated theorem proving. It's easy: you write the theorem you want proven and hope you get a proof. It works or it doesn't. Nice and benchmarkable.

Automated axiom creation seems a lot harder. How is an LLM supposed to know that "between any two points there is a line" formalizes an important property of physical space? Or how to suggest an alternative to Turing machines / lambda calculus that expresses the same underlying idea?

  • kvemkon 3 days ago

    > What we need is automated theorem discovery.

    I've been thinking mathematicians have fun doing math, making discoveries, crafting proofs.

    Does Tour de France & Co. make no sense since small, lightweight and powerful e-bicycles appeared?

    Using computer as a helper like bicycles is one thing, using LLMs seems more like e-bicycle and is something another.

  • ComplexSystems 3 days ago

    > We've had automated theorem proving since the 60s.

    By that logic, we've had LLMs since the 60s!

    > What we need is automated theorem discovery.

    I don't see any reason you couldn't train a model to do this. You'd have to focus it on generating follow-up questions to ask after reading a corpus of literature, playing around with some toy examples in Python and making a conjecture out of it. This seems much easier than training it to actually complete an entire proof.

    > Erdős discovered these theorems even if he wasn't really able to prove them. Euler and Gauss discovered a ton of stuff they couldn't prove. It is weird that nobody considers this to be intelligence.

    Who says they don't? I wouldn't be surprised if HarmonicMath, DeepMind, etc have also thought about this kind of thing.

    > Automated axiom creation seems a lot harder. How is an LLM supposed to know that "between any two points there is a line" formalizes an important property of physical space?

    That's a good question! It would be interesting to see if this is an emergent property of multimodal LLMs trained specifically on this kind of thing. You would need mathematical reasoning, visual information and language encoded into some shared embedding space where similar things are mapped right next to each other geometrically.