Comment by puttycat
Seems very incremental and very far from the pompous 'superintelligence' goal.
Seems very incremental and very far from the pompous 'superintelligence' goal.
If you can collapse "retrieve this complex chunk when it is needed" into a single token, what else can you put into a token?
"Send this through the math coprocessor." "Validate against the checklist." "Call out to an agent for X." "Recheck against input stream Y." And so on.
Retrieval augmentation is only one of many uses for this. If this winds up with better integration with agents, it is very possible that the whole is more than the sum of its parts.
It’s unlikely that the existing LLM architecture will evolve into anything that resembles superintelligence any more than it does already.
Which means that modifications to the architecture, and combining it with other components and approaches, are the next likely step. This paper fits that.