Comment by westurner
> Starting with theory might be the simplest way to explain,
Brilliant's AI course has step-by-step interactive textgen LLMs trained on TS (Swift) lyrics and terms of services with quizzes for comprehension and gamified points.
Here's a quick take:
LLM AI are really good at generating bytes that are similar to other bytes, but aren't yet very good at caring whether what they've generated is wrong or incorrect. Reinforcement Learning is one way to help prevent that.
AI Agents are built on LLMs. An LLM (Large Language Model) is a trained graph of token transition probabilities (a "Neural Network" (NN), a learning computer (Terminator (1984))). LLMs are graphical models. Clean your room. The grass is green and the sky is blue. Clean it well
AI Agents fail where LLMs fail at "accuracy" due to hallucinations even given human-curates training data.
There are lots of new methods for AI Agents built on LLMs which build on "Chain of Thought"; basically feeding the output from the model back through as an input a bunch of times. ("feed-forward")
But if you've ever heard a microphone that's too close to a speaker, you're already familiar with runaway feedback loops that need intervention.
There are not as many new Agentic AIs built on logical reasoning and inference. There are not as many AI Agents built on the Scientific Method that we know to be crucial to safety and QA in engineering.