Comment by someone3876

Comment by someone3876 4 days ago

1 reply

Have you actually used Gemini? I use it a lot for translation, and its context window is more like 150k tokens, rather than the 2M context window they say it has.

Der_Einzige 4 days ago

Be that as it may, long context window models which are good are not a mirage. By say late 2027, when the LLM providers figure out that they're using the wrong samplers, they will figure out how to get you 2 million output tokens per LLM call which stay coherent.