Readit News logoReadit News
jcheng commented on AGENTS.md outperforms skills in our agent evals   vercel.com/blog/agents-md... · Posted by u/maximedupre
w10-1 · 14 days ago
The key finding is that "compression" of doc pointers works.

It's barely readable to humans, but directly and efficiently relevant to LLM's (direct reference -> referent, without language verbiage).

This suggests some (compressed) index format that is always loaded into context will replace heuristics around agents.md/claude.md/skills.md.

So I would bet this year we get some normalization of both the indexes and the referenced documentation (esp. matching terms).

Possibly also a side issue: API's could repurpose their test suites as validation to compare LLM performance of code tasks.

LLM's create huge adoption waves. Libraries/API's will have to learn to surf them or be limited to usage by humans.

jcheng · 14 days ago
Would’ve been perfectly readable and no larger if they had used newline instead of pipe.
jcheng commented on Don't fall into the anti-AI hype   antirez.com/news/158... · Posted by u/todsacerdoti
tymscar · a month ago
I agree, however there you don’t compare technical knowledge alone, you also compare managerial skills.

With LLMs its admittedly a bit closer to doing it yourself because the feedback loop is much tighter

jcheng · a month ago
Yeah... and besides managerial skills, also product (using the word loosely) sense, user empathy, clarity of vision, communication skills. They've always been multipliers for programmers, even more so in this moment.
jcheng commented on Don't fall into the anti-AI hype   antirez.com/news/158... · Posted by u/todsacerdoti
tymscar · a month ago
I think you are right in saying that there is some deep intuition that takes months, if not years, to hone about current models, however, the intuition some who did nothing but talk and use LLMs nonstop two years ago would be just as good today as someone who started from scratch, if not worse because of antipatterns that don’t apply anymore, such as always starting a new chat and never using a CLI because of context drift.

Also, Simon, with all due respect, and I mean it, I genuinely look in awe at the amount of posts you have on your blog and your dedication, but it’s clear to anyone that the projects you created and launched before 2022 far exceed anything you’ve done since. And I will be the first to say that I don’t think that’s because of LLMs not being able to help you. But I do think it’s because what makes you really, really good at engineering you kept replacing slowly but surely with LLMs more and more by the month.

If I look at Django, I can clearly see your intelligence, passion, and expertise there. Do you feel that any of the projects you’ve written since LLMs are the main thing you focus on are similar?

Think about it this way: 100% of you wins against 100% of me any day. 100% of Claude running on your computer is the same as 100% of Claude running on mine. 95% of Claude and 5% of you, while still better than me (and your average Joe), is nowhere near the same jump from 95% Claude and 5% me.

I do worry when I see great programmers like you diluting their work.

jcheng · a month ago
> 95% of Claude and 5% of you, while still better than me (and your average Joe), is nowhere near the same jump from 95% Claude and 5% me.

I see what you're saying, but I'm not sure it is true. Take simonw and tymscar, put them each in charge of a team of 19 engineers (of identical capabilities). Is the result "nowhere near the same jump" as simonw vs. tymscar alone? I think it's potentially a much bigger jump, if there are differences in who has better ideas and not just who can code the fastest.

jcheng commented on Don't fall into the anti-AI hype   antirez.com/news/158... · Posted by u/todsacerdoti
embedding-shape · a month ago
> But what was the fire inside you, when you coded till night to see your project working? It was building.

I feel like this is not the same for everyone. For some people, the "fire" is literally about "I control a computer", for others "I'm solving a problem for others", and yet for others "I made something that made others smile/cry/feel emotions" and so on.

I think there is a section of programmer who actually do like the actual typing of letters, numbers and special characters into a computer, and for them, I understand LLMs remove the fun part. For me, I initially got into programming because I wanted to ruin other people's websites, then I figured out I needed to know how to build websites first, then I found it more fun to create and share what I've done with others, and they tell me what they think of it. That's my "fire". But I've met so many people who doesn't care an iota about sharing what they built with others, it matters nothing to them.

I guess the conclusion is, not all programmers program for the same reason, for some of us, LLMs helps a lot, and makes things even more fun. For others, LLMs remove the core part of what makes programming fun for them. Hence we get this constant back and forth of "Can't believe others can work like this!" vs "I can't believe others aren't working like this!", but both sides seems to completely miss the other side.

jcheng · a month ago
> For others, LLMs remove the core part of what makes programming fun for them.

Anecdotally, I’ve had a few coworkers go from putting themselves firmly in this category to saying “this is the most fun I’ve ever had in my career” in the last two months. The recent improvement in models and coding agents (Claude Code with Opus 4.5 in our case) is changing a lot of minds.

jcheng commented on Anthropic blocks third-party use of Claude Code subscriptions   github.com/anomalyco/open... · Posted by u/sergiotapia
dboon · a month ago
This is an unusual L for Anthropic. The unfortunate truth is that the engineering in opencode is so far ahead of Claude Code. Obviously, CC is a great tool, but that's more about the magic of the model than the engineering of the CLI.

The opencode team[^1][^2] built an entire custom TUI backend that supports a good subset of HTML/CSS and the TypeScript ecosystem (i.e. not tied to Opencode, a generic TUI renderer). Then, they built the product as a client/server, so you can use the agent part of it for whatever you want, separate from the TUI. And THEN, since they implemented the TUI as a generic client, they could also build a web view and desktop view over the same server.

It also doesn't flicker at 30 FPS whenever it spawns a subagent.

That's just the tip of the iceberg. There are so many QoL features in opencode that put CC to shame. Again, CC is a magical tool, but the actual nuts and bolts engineering of it is pretty damning for "LLMs will write all of our code soon". I'm sorry, but I'm a decent-systems-programmer-but-terminal-moron and I cranked out a raymarched 3D renderer in the terminal for a Claude Wrapped[^] in a weekend that...doesn't flicker. I don't mean that in a look-at-me way. I mean that in a "a mid-tier systems programmer isn't making these mistakes" kind of way.

Anyway, this is embarrassing for Anthropic. I get that opencode shouldn't have been authenticating this way. I'm not saying what they are doing is a rug pull, or immoral. But there's a reason people use this tool instead of your first party one. Maybe let those world class systems designers who created the runtime that powers opencode get their hands on your TUI before nicking something that is an objectively better product.

[^1] https://github.com/anomalyco/opentui

[^2] From my loose following of the development, not a monolith, and the person mostly responsible for the TUI framework is https://x.com/kmdrfx

[^3] https://spader.zone/wrapped/

jcheng · a month ago
An engineer on my team who is working on TUI stuff said that avoiding the flicker is difficult without affecting the ability to copy/paste using the mouse (something to do with "alternate screen mode"). I haven't used OpenCode (yet) but Google does turn up some questions (and suggested workarounds) around copy/paste.
jcheng commented on Anthropic blocks third-party use of Claude Code subscriptions   github.com/anomalyco/open... · Posted by u/sergiotapia
matt-p · a month ago
I think in fairness to anthropic they are winning in llms right? Since 3.7 they have been better than any other lab.
jcheng · a month ago
Arguably since 3.5, at least for coding and tool calling
jcheng commented on Nvidia's $20B antitrust loophole   ossa-ma.github.io/blog/gr... · Posted by u/ossa-ma
spiantino · 2 months ago
i thought so at first, but I did some digging and changed my mind. it's possible the following is how it goes:

- secondary transaction with the preferred shareholders (VCs) at some price that implies a 20b valuation

- founders quit and get new employment agreements

- some cash is transferred to the company as a license fee

- no acquisition means no DOJ approval

in this scenario the headline can be $20b but the cash expense can be much lower, you have full flexibility to direct whatever cash or equity you want to founders vs the rest of the company, as an up front payment or as retention/salary, and the founders have no hinderance from working on anything they touched at previous company because of IP license.

I actually bet this is how it went down. This is becoming the standard in the industry and it's just awful for the future of SV

jcheng · 2 months ago
To make this right they’d just have to amend the first part to “secondary transaction with shareholders at some price that implies a 20b valuation”.

Has there been any evidence yet that the VCs got paid for their shares but the left behind employees didn’t?

jcheng commented on A Proclamation Regarding the Restoration of the Dash   blog.nawaz.org/posts/2025... · Posted by u/BeetleB
jonathaneunice · 2 months ago
Cosigned!

Em dash forever! Along with en dash for numerical ranges, true ellipsis not that three-period crap, true typographic quotes, and all the trimmings! Good typography whenever and wherever possible!

jcheng · 2 months ago
And text figures! And proper small caps!!
jcheng commented on Ask HN: How can I get better at using AI for programming?    · Posted by u/lemonlime227
cerved · 2 months ago
So it does a forced reset of the dirt after each bash command? Does it confuse Claude? I frequently find it lacks path awareness of what it's working directory is
jcheng · 2 months ago
That’s exactly what it does, I’ve found it completely un-confuses Claude Sonnet 4.5.
jcheng commented on Ask HN: How can I get better at using AI for programming?    · Posted by u/lemonlime227
whatever1 · 2 months ago
For me what vastly improved the usefulness when working with big json responses was to install jq in my system and tell the llm to use jq to explore the json, instead of just trying to ingest it all together. For other things I explicitly ask it to write a script to achieve something instead of doing it directly.
jcheng · 2 months ago
That makes so much sense, it would make a great MCP. Maybe something similar for DOM manipulation; extracting text out of big, noisy HTML pages using a combination of Find Text with selector return values, and a DSL for picking and filtering DOM trees.

u/jcheng

KarmaCake day1235May 19, 2010
About
work email: joe AT posit DOT co
View Original