I assume thinking medium and hard would consume more tokens.
I feel the timing is bad for this release especially when deepseek R1 is still peaking. People will compare and might get disappointed with this model.
My vibe question checking suggests otherwise. Even o3-mini-high is not as good as r1, even though it's faster than r1. Considering o3-mini is more expensive per token. It's not clear o3-mini-high is cheaper than r1 either even r1 probably consumes more token per answer.
$1.1/m for input
$4.4/m for output
I assume thinking medium and hard would consume more tokens.
I feel the timing is bad for this release especially when deepseek R1 is still peaking. People will compare and might get disappointed with this model.