Comment by kesor
This is just like vibe coding. In vibe coding, you snapshot the results of the LLM's implementation into files that you reuse later.
This project could use something like that. Perhaps ask the LLM to implement a way to store/cache the snapshots of its previous answers. That way, the more you use it, the faster it becomes.