Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I use Claude Opus 4.6 as an enterprise user, and have also noticed a lobotomization. In recent weeks it's been much more self-correcting even within singular responses ("This is the problem - no wait, we already proved it can't be this - but actually ...") I'm wary of 4.7 being a change in this pattern, it's frustrating to have such a substantial change in experience every few months.


>..."This is the problem - no wait, we already proved it can't be this - but actually ..."

Ditto. Has me wondering why there isn't a reconciliation pass somewhere on the final output.

At least it's a decent signal for when model confidence is low.


Frustrating that the experience changes, and then they retire the better older model because it costs more, although it was better for everyone. The new ones are just geared better towards beating the benchmarks at a cheaper cost!




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: