Comment by DanHulton
You can build this today exactly as efficiently as you can when inference is 1000x faster, because the only things you can build with this is things that absolutely don't matter. The first bored high schooler who realizes that there's an LLM between them and the database is going to WRECK you.
this assumes the application is hosted as SaaS, but if the application makes sense as a personal/"desktop" app, that likely wouldn't matter.