If you’ve used either — or ideally both — would love to hear your insights. Feel answers to the following questions will provide some context when you respond:
- What are the strengths & weaknesses of each from your experience?
- Any tips/tricks or prompting techniques you use to get the most from these models?
- How do you typically use them? (via. native apps like ChatGPT, Claude; or via Cursor, GitHub Copilot, etc.?)
- What programming language(s) do you primarily use them with?
Hopefully this thread provides a useful summary and some additional tips for readers.
(I’ll start with mine in the comments)
- Sonnet 3.5 seems good with code generation and o1-preview seems good with debugging
- Sonnet 3.5 struggles with long contexts whereas o1-preview seems good at identifying interdependencies between files in code repo in answering complex questions
- Breaking the problem into small steps seems to yield better results with Sonnet
- I’m using primarily in Cursor/GH Copilot and with Python
https://github.com/StephanSchmidt/ai-coding-comparison/
(no comparison there yet, just some code to play around with)
Sonnet 3.5 if you can provide context (e.g. with cursor)
gpt-4o for UI design. Also for solving screenshots of interviews