Don't really agree, in my experience the switching context is extremely costly. I personally have trouble having even a couple of sessions running in parallel,Especially when I'm talking difficult hard to solve problems. Of course it's easy for trivial jobs, but it's not always the case. I have been much more successful in making my time worth by taking a look at the model's output and actively participating.It gives me time to think as well.When I have a list of simple tasks I just tell it to the model and it executes one after another.
rybosworld [3 hidden]5 mins ago
There's a lot more "telling" than "showing" going on.
By that I mean - the people claiming hyper-productivity from their GasTown setup never have actual products to demo.
hirako2000 [3 hidden]5 mins ago
Perhaps they earn $500k and worry spending any less than $250k in token may raise suspicion.
mcintyre1994 [3 hidden]5 mins ago
Something would be deeply wrong!
lanyard-textile [3 hidden]5 mins ago
I can do two or three at a time. I treat them a bit like queues: Last in first out, sort of like we do with our human peers.
We delegate work, we tend to some other work, and we code review much later in the day.
The secret to this mindset is that it doesn't always have to line up. Let your agent wait for you; You'll get to their output next.
EdNutting [3 hidden]5 mins ago
“Don’t pay attention to what Claude is doing, just spam your way through code and commands and hope nothing went wrong and you catch any code issues in review afterwards” is what this sounds like.
I will run parallel Claude sessions when I have a related cluster of bugs which can be fixed in parallel and all share similar context / mental state (yet are sufficiently distinct not to just do in one session with subagents).
Beyond that, parallel sessions to maybe explore some stuff but only one which is writing code or running commands that need checking (for trust / safety / security reasons).
Any waiting time is spent planning next steps (eg writing text files with prompts for future tasks) or reviewing what Claude previously did and writing up lists (usually long ones) of stuff to improve (sometimes with drafts prompts or notes of gotchas that Claude tripped up on the first time which I can prompt around in future).
Spend time thinking, not just motoring your way through tokens.
bwestergard [3 hidden]5 mins ago
If you are letting Claude run for seven minutes at a time, you aren't thinking hard enough about what you're building.
If you start trying to juggle multiple agents, you are doubling down on the wrong strategy.
Why should Claude finish complex tasks in less than seven minutes?
bwestergard [3 hidden]5 mins ago
The need for "complex tasks" should be exceptional enough that you're not building your workflow around them. A good example of such an exception would be kickstarting a port of a project for which you have a great test suite from one language to another. This is rare in most professional settings.
skydhash [3 hidden]5 mins ago
Computers are fast. If a physic engine can compute a game world in 1/60 of a second. The majority of the tasks should be done in less than 7 minutes.
Whenever I see transcript of a long running task, I see a lot of drifting of the agent due to not having any context (or the codebase is not organized) and it trying various way to gather information. Then it settle on the wrong info and produce bad results.
Greppability of the codebase helps. So do following patterns and good naming. A quick overview of the codebase and convention description also shortens the reflection steps. Adding helper tools (scripts) help too.
phainopepla2 [3 hidden]5 mins ago
Just show us the prompt you used to produce this post instead of the output
oidar [3 hidden]5 mins ago
I disagree with this take. I get that LLM produced text is filled with crappy, over the top writing in pretty much all cases, but if a prompter/writer/blogger is using it iteratively, the LLM output is going to be way better than their writing. Also, if a person is using LLMs to write articles, do you really want to see their likely even worse writing?
the_af [3 hidden]5 mins ago
Nice catch. Look at this at the end:
> jc is open source. If you have improvements, have your Claude open a PR against mine. I don’t accept human-authored code.
So it seems not only does the author reject human-authored PRs, they also refuse human-authored blog posts.
snovv_crash [3 hidden]5 mins ago
I wonder if they also only want agents to read it, not people.
oytis [3 hidden]5 mins ago
When computer works, it's sword fighting time. I don't make the rules
And for those who are feeling smug, that last one (which I still consider fairly recent) was 14 years ago
armSixtyFour [3 hidden]5 mins ago
I'm not slacking off, Claude is shmorgalizing.
kubb [3 hidden]5 mins ago
Wasn't there some recent discovery that context switching is harmful to your brain?
cruffle_duffle [3 hidden]5 mins ago
This advice will be very dated when inference gets an order of magnitude faster. And it will happen—it’s classic tech. Probably will even follow moores law or something.
Wait until that 8 minute inference is only a handful of seconds and that is when things get real wild and crazy. Because if the time inference takes isn’t a bottleneck… then iteration is cheap.
trjordan [3 hidden]5 mins ago
I'd offer a different approach: think about how you're going to validate. An only-slightly-paraphrased Claude conversation I had yesterday:
> me: I want our agent to know how to invoke skills.
> Claude: [...]
> Claude: Done. That's the whole change. No MCP config, no new env vars, no caller changes needed.
> me: ok, test it.
> Claude: This is a big undertaking.
That's the hard part, right? Maybe Claude will come back with questions, or you'll have to kick it a few times. But eventually, it'll declare "I fixed the bug!" or summarize that the feature is implemented. Then what?
I get a ton of leverage figuring this out what I need to see to trust the code. I work on that. Figure out if there's a script you can write that'll exercise everything and give you feedback (2nd claude session!). Set up your dev env so playwright will Just Work and you can ask Claude to click around and give you screenshots of it all working. Grep a bunch and make yourself a list of stuff to review, to make sure it didn't miss anything.
BeetleB [3 hidden]5 mins ago
Very painful to read.
jolt42 [3 hidden]5 mins ago
Yes. I agree with the problem statement, but have no idea what the solution is BUT it does involve lots of key bindings.
tkzed49 [3 hidden]5 mins ago
was this written using a LinkedIn skill
servercobra [3 hidden]5 mins ago
This looks absolutely wonderful. Is it possible to run against Claude remotely (e.g. on a VM?). Or should I ask Claude to add that?
mbo [3 hidden]5 mins ago
I'm not sure I'm understanding this workflow. Perhaps a small tutorial / walkthrough hosted on YouTube or asciinema might help people understand.
hirako2000 [3 hidden]5 mins ago
It's just a process to loop over a number of cycle prompting each thing taking minutes to run. It's a recipe for a massive headache as context switching costs more than 7 minutes (that arbitrary number the article came up with)
teaearlgraycold [3 hidden]5 mins ago
> The fix is obvious: work on something else while Claude runs.
Disagree. The fix is actually counter-intuitive: give Claude smaller tasks so that it completes them in less time and you remain in the driver's seat.
flakiness [3 hidden]5 mins ago
The old saying is "don't multitask" but apparently that time is gone.
I wonder what people think about this. I know there is a class of SWE/dev who now consider oneself as "the manager of agents". Good luck to them and articles like this would work for these people.
I'm not there yet and I hope I don't have to. I'm not a LLM and my mental model is (I believe) more than a markdown. But I haven't figured out the mental model that works for me, still staring at the terminal Claude blinking the cursor, sticking to "don't multitask" dogma.
chis [3 hidden]5 mins ago
Hackernews needs to nominate an elite crew of individuals who can tell when an article is AI slop and flag it.
jedberg [3 hidden]5 mins ago
Or, wait and take a little break so you don't burn out. I miss the days where you had to wait for code to compile or for your "big data" job to run, so you could give yourself a little mini break.
By that I mean - the people claiming hyper-productivity from their GasTown setup never have actual products to demo.
We delegate work, we tend to some other work, and we code review much later in the day.
The secret to this mindset is that it doesn't always have to line up. Let your agent wait for you; You'll get to their output next.
I will run parallel Claude sessions when I have a related cluster of bugs which can be fixed in parallel and all share similar context / mental state (yet are sufficiently distinct not to just do in one session with subagents).
Beyond that, parallel sessions to maybe explore some stuff but only one which is writing code or running commands that need checking (for trust / safety / security reasons).
Any waiting time is spent planning next steps (eg writing text files with prompts for future tasks) or reviewing what Claude previously did and writing up lists (usually long ones) of stuff to improve (sometimes with drafts prompts or notes of gotchas that Claude tripped up on the first time which I can prompt around in future).
Spend time thinking, not just motoring your way through tokens.
If you start trying to juggle multiple agents, you are doubling down on the wrong strategy.
https://hbr.org/2010/12/you-cant-multi-task-so-stop-tr
Whenever I see transcript of a long running task, I see a lot of drifting of the agent due to not having any context (or the codebase is not organized) and it trying various way to gather information. Then it settle on the wrong info and produce bad results.
Greppability of the codebase helps. So do following patterns and good naming. A quick overview of the codebase and convention description also shortens the reflection steps. Adding helper tools (scripts) help too.
> jc is open source. If you have improvements, have your Claude open a PR against mine. I don’t accept human-authored code.
So it seems not only does the author reject human-authored PRs, they also refuse human-authored blog posts.
https://xkcd.com/1053/
And for those who are feeling smug, that last one (which I still consider fairly recent) was 14 years ago
Wait until that 8 minute inference is only a handful of seconds and that is when things get real wild and crazy. Because if the time inference takes isn’t a bottleneck… then iteration is cheap.
> me: I want our agent to know how to invoke skills.
> Claude: [...]
> Claude: Done. That's the whole change. No MCP config, no new env vars, no caller changes needed.
> me: ok, test it.
> Claude: This is a big undertaking.
That's the hard part, right? Maybe Claude will come back with questions, or you'll have to kick it a few times. But eventually, it'll declare "I fixed the bug!" or summarize that the feature is implemented. Then what?
I get a ton of leverage figuring this out what I need to see to trust the code. I work on that. Figure out if there's a script you can write that'll exercise everything and give you feedback (2nd claude session!). Set up your dev env so playwright will Just Work and you can ask Claude to click around and give you screenshots of it all working. Grep a bunch and make yourself a list of stuff to review, to make sure it didn't miss anything.
Disagree. The fix is actually counter-intuitive: give Claude smaller tasks so that it completes them in less time and you remain in the driver's seat.
I wonder what people think about this. I know there is a class of SWE/dev who now consider oneself as "the manager of agents". Good luck to them and articles like this would work for these people.
I'm not there yet and I hope I don't have to. I'm not a LLM and my mental model is (I believe) more than a markdown. But I haven't figured out the mental model that works for me, still staring at the terminal Claude blinking the cursor, sticking to "don't multitask" dogma.
Of course there is a relevant XKCD: https://xkcd.com/303/
Is this sarcasm? If not, I wonder why.