Q: Output Context Window
Dear Qolaba Team,
I am interested in your product, but I did some tests (on the free plan) and it seems that there is a 5k output context window limitation.
Is this also true on the LTD? I am interested in using Gem 2.5 pro 2 mio context window to produce long reports.
Thank you very much in advance.
Best regards,
John

Annapurna_Qolaba
Jun 9, 2025A: Hi John, great question!
Yes, this is true for all plans, including the LTD.
The total context window (input + output combined) for models like Gemini 2.5 Pro is up to 1 million tokens. However, the maximum output per single response is around 6,000 tokens.
For longer outputs, you can simply continue with a follow up prompt like “continue”, and the model will pick up right where it left off. So yes, you can absolutely produce long reports over multiple prompts while staying within the full 1M context range.
Hope that helps clarify!

Verified purchaser
Hi thank you very much.
Is there any plan to allow a bigger output context window in the future ?
Also how does it compare to straico ? Are you going to develop similar “unifying”, “prompt assistant”, “models combos”, etc. Capabilities?
Thanks in advance
Best regards
We’re always keeping an eye on improvements released by model providers like Google, OpenAI, and Anthropic. Currently, output limits are tied to their API constraints (e.g., Gemini 2.5 Pro caps single responses at ~6,000 tokens). But as models evolve to allow larger outputs, we’ll be quick to update Qolaba to support them.
On Straico-style features: Yes, we’re already working on agent flows, custom instructions, model switching, etc, all in one space. Our aim is to make Qolaba your unified AI workspace, and you’ll start seeing more features roll out gradually.