Comment by gosub100

Comment by gosub100 3 days ago

3 replies

Forgive my ignorance about AI, but had anyone tried a "nondeterministic" language that somehow uses learning to approximate the answer? I'm not talking about the current cycles where you train your model on a zillions of inputs, tune it, and release it. I mean a language where you tell it what a valid output looks like, and deploy it. And let it learn as it runs.

Ex: my car's heater doesn't work the moment you turn it on. So if I enter the car one of my first tasks is to turn the blower down to 0 until the motor warms up. A learning language could be used here, given free reign over all the (non-safety-critical) controls, and told that it's job is to minimize the number of "corrections" made by the user. Eventually it's reward would be gained by initializing the fan blower to 0, but it might take 100 cycles to learn this. Rather that train it on a GPU, a language could express the reward and allow it to learn over time, even though it's output would be "wrong" quite often.

That's an esoteric language I'd like to see.

abhgh 2 days ago

Wouldn't this be an optimization problem, that's to say, something like z3 should be able to do - [1], [2]?

I was about to suggest probabilistic programming, e.g., PyMC [3], as well, but it looks like you want the optimization to occur autonomously after you've specified the problem - which is different from the program drawing insights from organically accumulated data.

[1] https://github.com/Z3Prover/z3?tab=readme-ov-file

[2] https://microsoft.github.io/z3guide/programming/Z3%20Python%...

[3] https://www.pymc.io/welcome.html