GTP 4.5 is not a reasoning model. Reasoning models outperform it clearly. Even OpenAIs o3-mini is smarter while being magnitudes cheaper. Those 2 should be compared in my opinion.
GPT 4.5 feels like a failed experiment to see how far you can push non-thinking models.
>GPT 4.5 feels like a failed experiment to see how far you can push non-thinking models
It's not a failed experiment, it's a very good experiment, because it produced a very useful piece of information for the world (that there's limited return to further size scaling).
Outperform in what way? Reasoning models may be able to solve problems correctly a bigger percentage of time, but they burn many tokens to get there. So they’re much less efficient, both in latency and ultimately environmental cost.
I’m quite late here, but if you want a diagram you can ask the LLM to output Mermaid syntax and then paste that into Excalidraw or something else that can render based on Mermaid.