Comment by gjm11
No specific critique here other than "it was written by an LLM and this seems worth pointing out given that LLMs are bad at actually understanding difficult mathematics".
(In a different comment I make some actual criticisms of what you wrote. I see you replied to my comment there, and that's a more appropriate place to discuss actual ideas. I don't see much point in criticizing LLM output in a field LLMs are bad at.)
Anyway: (1) no, it wasn't clear. I wouldn't generally take "I nerd-sniped myself. Here's a more fleshed-out sketch of ..." to mean "Here's something written for me by an LLM". I'd take it to imply that the person had done the fleshing-out themself. And (2) no, the problem wasn't that you used words I don't understand. It's certainly possible that your ideas are excellent and I just don't understand them, but I'm a mathematician myself and none of the words scare me.
"no, it wasn't clear. I wouldn't generally take "I nerd-sniped myself. Here's a more fleshed-out sketch of ..." to mean "Here's something written for me by an LLM". I'd take it to imply that the person had done the fleshing-out themself."
aaaaaaaah, I think you finally helped me notice something subtle in the way I use LLMs than other people. It sounds obvious now that I think about it but I never considered people use LLMs like google whereas I use it more like a real time thought transcriber (e.g. Dragon Naturally Speaking but not shite :P) Since it's trained on a RAG based off of my own polished thoughts, I've set it up as a meta-circular evaluator to do linguistic filtration (basically Fourier kernels on clip embedding space that map to various measures of "conceptual clarity").
So the LLM-ness of it to me is a clear-flag that this is hastily dictated.