Comment by nitinram

Comment by nitinram 20 hours ago

1 reply

This is super cool! I attempted to use this on a project and kept running into "This model's maximum context length is 200000 tokens. However, your messages resulted in 459974 tokens. Please reduce the length of the messages." I used open ai o4-mini. Is there an easy way to handle this gracefully? Basically if you had thoughts on how to make some tutorials for really large codebases or project directories?

zh2408 19 hours ago

Could you try to use gemini 2.5 pro? It's free every day for first 25 requests, and can handle 1M input tokens