LLM Coding Comparisons
Users compare the performance of AI models like Claude 3.5 Sonnet, GPT-4o, o1, and Gemini, especially for coding tasks, sharing experiences, benchmarks, and preferences.
Activity Over Time
Top Contributors
Keywords
Sample Comments
Is there any evidence this works better than Claude 3.5?
Because there's less drama? I use Claude 3.5 Sonnet every day for helping me with coding. It seems to just work. It's been much better than GPT-4 for me, haven't tried o1, but don't really feel the need, very happy with Claude.
Did you compare the performance with o1 or Claude 3.5 Sonnet?
Has this been benchmarked or something? How does it compare to chatgpt3.5?
Have you tried Claude (3.5 Sonnet)?
Which ChatGPT model have you been using? In my experience nothing beats 4. (Not claude, not 4o)
In my experience it's better than GPT3.5, not as good as GPT4.
Those who use these can you compare the quality of code compared to Claude Sonnet 4.5 or Opus 4.1?
Claude Opus (largest v3 model) consistently outperforms GPT-4 for me. Better at following prompts, _feels_ much better.
I am the opposite. Find GPT 5.2 much worse. Sticking only with gemini and claude.