It would be interesting to swap out Ollama for LM Studio and use their built-in MLX support and see the difference.
[1] https://iomstamps.com/collections/wakeman [2] https://www.bbc.co.uk/news/articles/clyqe679gqno
Seems to be just a rehashing of the same idea but instead of XML, and diagrams, it's now some free-text to be interpreted by LLMs, so much less deterministic and will probably fail just like UML failed.
People also tend to forget about Peter Naur's take on "Programming as Theory Building" [0], the program is, in itself, the theory of what's implemented. A spec cannot replace that.
Most of the items in this list fall prey to it, but it is maybe best exemplified by this one:
> A writing app that lets you “request a critique” from a bunch of famous writers. What would Hemingway say about this blog post? What did he find confusing? What did he like?
Any app that ever claimed to tell you what "Hemingway would say about this blog post" would evidently be lying — it'd be giving you what that specific AI model generates in response to such a prompt. 100 models would give you 100 answers, and none of them could claim to actually "say what Hemingway would've said". It's not as if Hemingway's entire personality and outlooks are losslessly encoded into the few hundreds of thousands of words of writing/speech transcripts we have from him, and can be reconstructed by a sufficiently beefy LLM.
So in effect it becomes an exercise of "can you fool the human into thinking this is a plausible thing Hemingway would've said".
The reason why you would care to hear Hemingway's thought on your writing, or Steve Jobs' thoughts on your UI design, is precisely because they are the flesh-and-bone, embodied versions of themselves. Anything else is like trying to eat a picture of a sandwich to satisfy your hunger.
There's something unsettling that so many people cannot seem to cut clearly through this illusion.
[1] https://arxiv.org/pdf/2509.18480 [2] https://github.com/apple/ml-simplefold
So in the end, it's code that I know very, very well. I could have written it but it would have taken me about 3x longer when all is said and done. Maybe longer. There are usually parts that have difficult functions but the inputs and outputs of those functions are testable so it doesn't matter so much that you know every detail of the implementation, as long as it is validated.
This is just not junior stuff.
Seniors can make this explicit to models and use them to automate "the code they would have written," whereas a junior doesn’t know what they would have written nor how they would have solved it absent a LLM.
Same applies to all fields: LLMs can be either huge leverage on top of existing knowledge or a crutch for a lack of understanding.
What gets me is that this isn't even a software moat anymore - it's literally just whoever can get their hands on enough GPUs and power infrastructure. TSMC and the power companies are the real kingmakers here. You can have all the talent in the world but if you can't get 100k H100s and a dedicated power plant, you're out.
Wonder how much of this $13B is just prepaying for compute vs actual opex. If it's mostly compute, we're watching something weird happen - like the privatization of Manhattan Project-scale infrastructure. Except instead of enriching uranium we're computing gradient descents lol
The wildest part is we might look back at this as cheap. GPT-4 training was what, $100M? GPT-5/Opus-4 class probably $1B+? At this rate GPT-7 will need its own sovereign wealth fund
From Dario’s interview on Cheeky Pint: https://podcasts.apple.com/gb/podcast/cheeky-pint/id18210553...