If you're still picking LLM providers by gut feeling, you're leaving money on the table. I ran 5 developer use cases through Claude 3.5 Sonnet, GPT-4o, and Gemini 2.0 Flash using PromptFuel to measure token usage and cost. The results?
More interesting than "fastest wins." Here's what I found. The Setup I took 5 tasks I actually do in PromptFuel development: JSON schema validation prompt — cat