Claude Opus 4.7 (anthropic.com)

by AlphaWeaver 12 comments 186 points
Read article View on HN

12 comments

[−] ChrisArchitect 29d ago
[−] tomhow 29d ago
Comments moved thither. Thanks!
[−] AlphaWeaver 29d ago
Might be better to update the URL to this, actually: https://www.anthropic.com/news/claude-opus-4-7
[−] jameson 29d ago
How should one compare benchmark results?

For example, SWE-bench Pro improved ~11% compared with Opus 4.6. Should one interpret it as 4.7 is able to solve more difficult problems? or 11% less hallucinations?

[−] constantius 29d ago
Not related to this release, but is anyone aware of what's happening with Deepseek? The usual cascade of synced releases has been lacking this frontier lab whale for a while now.
[−] rvz 29d ago

> Not related to this release, but is anyone aware of what's happening with Deepseek?

Given that no-one is talking about DeepSeek, I assume it is coming this month.

They are still releasing research papers and that is what really matters and not the .1 increment releases of AI models to massage benchmarks or create hype around.

[−] cmrdporcupine 29d ago
There's been months of "DeepSeek v4 next week!" rumours and none have panned out.

They're either stuck/dead or they're sitting on something really fantastic that they only want to release once they've perfected it.

My realistic side thinks the former, my optimism on the latter.

In the meantime, GLM 5.1 is actually really good.

[−] constantius 19d ago
You were right to be optimistic.
[−] bsaul 29d ago
i tried to find an API pricing for GLM 5.1 but couldn't find any on the homepage. How are you using it ?
[−] cmrdporcupine 29d ago
per-token via DeepInfra, who hosts it as one of their models.

https://deepinfra.com/zai-org/GLM-5.1

[−] grandinquistor 29d ago
Quite a big improvement in coding benchmarks, doesn’t seem like progress is plateauing as some people predicted.
[−] hansmayer 29d ago
Ah, here we go again.
[−] vomayank 29d ago
[flagged]
[−] pukaworks 28d ago
[dead]