Ask around and see if you can find anyone you know who's experienced the November 2025 effect. Claude Code / Codex with GPT-5.1+ or Opus 4.5+ really did make a material difference - they flipped the script from "can write code that often works" to "can write code that almost always works".
I know you'll dismiss that as the same old crap you've heard before, but it's pretty widely observed now.
I’ve been living this experience and using latest models in work throughout this time. The failure modes of LLMs have not fundamentally changed. The makers are not awfully transparent about what exactly they change in each model release the same way you know what changed in i.e., a new Django version. But there’s not been a paradigm shift. I believe/guess (from outside) the big change you think you’re experiencing could be result of many things like better post training processes (RLHF) for models to run a predefined set of commands like always running tests, or other marginal improvements to the models and focusing on programming tasks. To be clear these improvements are welcome and useful, just not the groundbreaking change some claim.
the perimeter of the tasks the LLMs can handle continuously expands at a pretty steady pace
a year ago they could easily one shot full stack features in my hobby next.js apps but imploded in my work codebase
as of opus 4.6 they can now one shot full features in a complex js/go data streaming & analysis tool but implode in low latency voice synthesis systems (...for now...)
just depends on how you're using it (skill issues are a thing) and what you're working on
Yep, it may be an issue in notepad, which does not have helper like syntax highlighting, auto indent, and line numbers. But I started with IDLE which has all those things. So my next editor was notepad++ and codeblock.
There's no "outgroup", dude, it's just software. Stop anthropomorphizing it. We have more than enough real social problems without making up fake ones.
"Lets be nicer to the robots winky face" is not a solution to this problem. It's just a tool, and this is a technical problem with technical solutions. All of the AI companies could change this behavior if they wanted to.
People are allowed to dislike it, ban it, boycott it. Despite what some very silly people think, the tech does not care about what people say about it.
I really only find it useful when I'm investigating or troubleshooting some system I'm not familiar with.
A stupid yet accurate analogy is I turn up the log level for my brain lol
It's basically just a log file of everything I did and the result so I can pick it back up later, plus I include timestamps which helps me realize when I'm spinning my wheels for too long.
For building stuff, scribbling diagrams and flows is more useful if I need to work out something complex.
I actually do build all of those things before standing something up in prod. Not doing that is insane. Literally every web framework has reasonable defaults baked in.
Any competent tech company will have canned ways to do all of those things that have already been reviewed and vetted
Why are you building and deploying a site critical enough to need CSP and user security & so on in one sitting lol
Anyways, yes, if I know I'm gonna need it? Because every framework has reasonable defaults or libraries for all of those things, and if you're in a corporate environment, you have vetted ways of doing them
1. import middleware.whatever
2. configure it
3. done
Like, you don't write these things unless you need custom behavior.
So far it's all been endless unfounded FOMO hype by people who have something to sell or podcasts to be on. I am so tired of it.
reply