Comment by TeMPOraL
Maybe the prompt you used was more Claude-friendly than Gemini-friendly?
I'm only half-joking. Different models process their prompts differently, sometimes markedly so; vendors document this, but hardly anyone pays any attention to it - everyone seems to be writing prompts for an idealized model (or for whichever one they use the most), and then rate different LLMs on how well they respond.
Example: Anthropic documents both the huge impact of giving the LLM a role in its system prompt, and of structuring your prompt with XML tags. The latter is, AFAIK, Anthropic-specific. Using it improves response quality (I've tested this myself), and yet as far I've seen, no BYOK tool offering multiple vendor support respects or leverages that.
Maybe Gemini has some magic prompt features, too? I don't know, I'm in the EU, and Google hates us.
Possibly. But my Claude prompts work fine on ChatGPT, the only difference being ChatGPT isn't very good. I pay for both.
I would not pay for Gemini - which is presumably why they've added it for "free" for everyone.
My anthropic prompts in the API are structured. I've got one amazing API prompt that has 67 instructions, and gives mind-blowing results (to the point that it has replaced a human) but for a simple question I don't find value in that. And, frankly, 'consumer'-facing AI chatbots shouldn't need prompting expertise for basic out of the box stuff.
The prompt I used in this example was simply "Please extract the data points contained within this report and present as structured data"
> and yet as far I've seen, no BYOK tool offering multiple vendor support respects or leverages that
When you say BYOK tool do you mean effectively a GUI front end on the API? I use typingmind for quickly throwing things at my API keys for testing, and I'm pretty sure you can have a persistent custom system prompt, though I think you'd need to input it for each vendor/model.