This might be why agentic development/vibe coding leads to more burn out. It's been a long time since I've truly been 'stuck' on a problem and needed to sleep on it to figure out the answer. Now I just ask Claude to fix it until it's fixed...
Sounds like polyphasic sleeping might re-emerge as the lifestyle solution. Instead of waiting for agents to complete, you should sleep on the response so when you arise you have the optimized prompt ready to go and a reset on your energy to prevent the burnout.
Amusingly this is an almost-exact description of how I work on my current project, sharc. I'm porting Arc to Common Lisp, and implementing as many HN features as I can. I've been documenting as I go with handoffs: https://github.com/shawwn/sharc/tree/main/docs/agents/handof... (Also thanks partly to dang, who is kind enough to find time to answer an email here and there about their current Arc stack.)
At one point I was working so hard that Claude actually suggested, all on its own, that I should get some sleep.
FWIW I've had the opposite experience. Whenever I work late the output is absolute garbage. If I work past midnight it takes me 3 hours to get done what would have taken me 30 mins in the morning, and with way less frustration and stress. Your inputs to the LLM are only as good as how fresh your mind is so I've made it a rule to not work past midnight (unless there's an emergency).
In the good old days you would reach flow and actually know when you're too tired to continue. Now you can just say "please just fix it" over and over again and get yourself in a slophole much easier.
Most software doesn't really have "hard enough problems" unless you're working in deep tech. The majority of SWEs are probably working on some sort of SaaS which isn't super challenging for a model like Opus 4.7. Most of the problems I face are on the product side, which I do need to take time to think through, but it's not as challenging as debugging in the good old days.
How do you go from SaaS to “not super challenging”? The part of a SaaS product that I’m working on uses graph algorithms to work with what’s essentially an interactive form. There’s some mildly university-level computer science stuff and it’s mixed with enough domain expertise that Opus 4.7 is still unable to make even small changes without breaking everything or going against the architecture.
Maybe the only solution to GPTisms is infinite context. If I'm talking to my coworker every day I would consciously recognize when I already used a metaphor recently and switch it up. However if my memory got reset every hour, I certainly might tell the same story or use the same metaphor over and over.
> However if my memory got reset every hour, I certainly might tell the same story or use the same metaphor over and over.
All people repeat the same stories and phraseology to some extent, and some people are as bad or worse than LLM chat bots in their predictability. I wonder if the latter have weak long-term memory on the scale of months to years, even if they remember things well from decades ago.
Honestly I think there is more to it - even with infinite context, the LLM needs some kind of intelligence to know what is noise and what is not, you resort to "thinking" - making it create garbage it then feeds to itself.
Learning a language is a big complex task, but it is far from real intelligence.
Lock in is pretty easy these days. Just a dummy example, Claude models are trained on their `str_replace_based_edit_tool` edit tool[1] which is very different from OpenAI's `apply_patch` tool[2].
Also Discord - tons of people use Discord as a social network and keep up with friends. I must have 5 friend groups that have their own Discords with some overlap.
So did you disclose this responsibly? Posting about it publicly first is asking for that sensitive data to be leaked. Might as well hack and repost that PII yourself.
This is not a data leakage.
They deliberately included 999 of their customers' email addresses in publicly accessible JavaScript code in order to test certain features on them.
Certainly that wasn't intentional to broadcast to the public? Sounds like a textbook data leak.
> A data leak is the unauthorized, often unintentional exposure of sensitive, confidential, or personal information to an external party, usually resulting from weak infrastructure, human error, or system errors.
Consider medical device software. Often embedded C code, needs to be rigorously documented and tested, has longer development cycles, and certainly no attitudes of "bugs are fine, ship it and we'll patch later."
Yes. High value work where cost (mostly) doesn't matter. For example, if I need to look over a legal doc for possible mistakes (part of a workflow i have), it doesn't matter (in my case) whether it costs $0.01 or $10.00, since it's a somewhat infrequent event. So i'll pay $9.99 more, even if the model is only slightly better.
I'm surprised I never heard people talking about using -Pro variants, even though their rates ($125-175/M?) aren't drastically larger than old Opus ($75/M), which people seemed to use
1. https://artificialanalysis.ai/models/grok-4-3
reply