Readit News logoReadit News
SatvikBeri commented on Speed up responses with fast mode   code.claude.com/docs/en/f... · Posted by u/surprisetalk
nick49488171 · 15 hours ago
For us mere mortals, how fast does a normal developer for through a MTok. How about a good power user?
SatvikBeri · 10 hours ago
I use one Claude instance at a time, roughly fulltime (writes 90% of my code). Generally making small changes, nothing weird. According to ccusage, I spend about $20 of tokens a day, a bit less than 1 MTOK output tokens a way. So the exact same workflow would be about $120 for higher speed.
SatvikBeri commented on My AI Adoption Journey   mitchellh.com/writing/my-... · Posted by u/anurag
fergie · 2 days ago
He explicitly said "I don't work for, invest in, or advise any AI companies." in the article.

But yes, Hashimoto is a high profile CEO/CTO who may well have an indirect, or near-future interest in talking up AI. HN articles extoling the productivity gains of Claude on HN do generally tend to be from older, managerial types (make of that what you will).

SatvikBeri · 2 days ago
What made me feel old today: seeing a 36-year-old referred to as an older type
SatvikBeri commented on My AI Adoption Journey   mitchellh.com/writing/my-... · Posted by u/anurag
tudelo · 2 days ago
First off, appreciate you sharing your perspective. I just have a few questions.

> I've gone back to managing the context window in Emacs because I can't be bothered to learn how to deal with another model family that will be thrown out in six months.

Can you expand more on what you mean by that? I'm a bit of a noob on llm enabled dev work. Do you mean that you will kick off new sessions and provide a context that you manage yourself instead of relying on a longer running session to keep relevant information?

> Unironically learning vim or Emacs and the standard Unix code tools is still the best thing you can do to level up your llm usage.

I appreciate your insight but I'm failing to understand how exactly knowing these tools increases performance of llms. Is it because you can more precisely direct them via prompts?

SatvikBeri · 2 days ago
One thing to keep in mind is that the core of an LLM is basically a (non-deterministic) stateless function that takes text as input, and gives text as output.

The chat and session interfaces obscure this, making it look more stateful than it is. But they mainly just send the whole chat so far back to the LLM to get the next response. That's why the context window grows as a chat/session continues. It's also why the answers tend to get worse with longer context windows – you're giving the LLM a lot more to sift through.

You can manage the context window manually instead. You'll potentially lose some efficiencies from prompt caching, but you can also keep your requests much smaller and more relevant, likely spending fewer tokens.

SatvikBeri commented on Claude Opus 4.6   anthropic.com/news/claude... · Posted by u/HellsMaddy
ifwinterco · 3 days ago
On benchmarks GPT 5.2 was roughly equivalent to Opus 4.5 but most people who've used both for SWE stuff would say that Opus 4.5 is/was noticeably better
SatvikBeri · 3 days ago
I pretty consistently heard people say Codex was much slower but produced better results, making it better for long-running work in the background, and worse for more interactive development.
SatvikBeri commented on GPT-5.3-Codex   openai.com/index/introduc... · Posted by u/meetpateltech
jahsome · 3 days ago
Another day, another hn thread of "this model changes everything" followed immediately by a reply stating "actually I have the literal opposite experience and find competitor's model is the best" repeated until it's time to start the next day's thread.
SatvikBeri · 3 days ago
I use Claude Code every day, and I'm not certain I could tell the difference between Opus 4.5 and Opus 4.0 if you gave me a blind test
SatvikBeri commented on VisualJJ – Jujutsu in Visual Studio Code   visualjj.com/... · Posted by u/demail
Zacharias030 · 6 days ago
do you use it with git worktrees? or how do you handle multiple claudes?
SatvikBeri · 6 days ago
I use jj workspaces, which are pretty similar to git worktrees. But 95% of the time I just use one Claude.
SatvikBeri commented on VisualJJ – Jujutsu in Visual Studio Code   visualjj.com/... · Posted by u/demail
stouset · 7 days ago
Claude Code has zero issue using jj, for what it’s worth.
SatvikBeri · 7 days ago
I noticed it get a lot better with the 4.5 models specifically. Before then, I would have to tell Claude "do a web search for the docs for jj, then do XYZ." Now I just tell it to "do XYZ using jj"
SatvikBeri commented on What I learned building an opinionated and minimal coding agent   mariozechner.at/posts/202... · Posted by u/SatvikBeri
jrm4 · 7 days ago
This is the first I'm hearing of this pi-agent thing and HOW DO PEOPLE TECH DECIDE TO NAME THINGS?

Seriously. Is creator not aware that "pi" absolutely invokes the name of another very important thing? sigh.

SatvikBeri · 7 days ago
From the article: "So what's an old guy yelling at Claudes going to do? He's going to write his own coding agent harness and give it a name that's entirely un-Google-able, so there will never be any users. Which means there will also never be any issues on the GitHub issue tracker. How hard can it be?"
SatvikBeri commented on What I learned building an opinionated and minimal coding agent   mariozechner.at/posts/202... · Posted by u/SatvikBeri
dagss · 7 days ago
> from copying and pasting code into ChatGPT, to Copilot auto-completions [...], to Cursor, and finally the new breed of coding agent harnesses like Claude Code, Codex, Amp, Droid, and opencode

Reading HN I feel a bit out of touch since I seem to be "stuck" on Cursor. Tried to make the jump further to Claude Code like everyone tells me to, but it just doesn't feel right...

It may be due to the size of my codebase -- I'm 6 months into solo developer bootstrap startup, so there isn't all that much there, and I can iterate very quickly with Cursor. And it's mostly SPA browser click-tested stuff. Comparatively it feels like Claude Code spends an eternity to do something.

(That said Cursor's UI does drive me crazy sometimes. In particular the extra layer of diff-review of AI changes (red/green) which is not integrated into git -- I would have preferred that to instead actively use something integrated in git (Staged vs Unstaged hunks). More important to have a good code review experience than to remember which changes I made vs which changes AI made..)

SatvikBeri · 7 days ago
Seems like there's a speed/autonomy spectrum where Cursor is the fastest, Codex is the best for long-running jobs, and Claude is somewhere in the middle.

Personally, I found Cursor to be too inaccurate to be useful (possibly because I use Julia, which is relatively obscure) – Opus has been roughly the right level for my "pair programming" workflow.

SatvikBeri commented on What I learned building an opinionated and minimal coding agent   mariozechner.at/posts/202... · Posted by u/SatvikBeri
SatvikBeri · 7 days ago
I particularly liked Mario's point about using tmux for long-running commands. I've found models to be very good at reading from / writing to tmux, so I'll do things like spin up a session with a REPL, use Claude to prototype something, then inspect it more deeply in the REPL.

u/SatvikBeri

KarmaCake day4282May 19, 2011
About
satvik.beri@gmail.com
View Original