Reason? Maybe. But there's one limitation that we currently have no idea how to overcome; LLMs don't know how much they know. If they tell you they don't something it may be a lie. If they tell you they do, this may be a lie too. I, a human, certainly know what I know and what I don't and can recall from where I know the information
> Mathematicians used to comb through model solutions because earlier systems would quietly flip an inequality or tuck in a wrong step, creating hallucinated answers.
> Brown says the updated IMO reasoning model now tends to say “I’m not sure” whenever it lacks a valid proof, which sharply cuts down on those hidden errors.
> TLDR, the model shows a clear shift away from hallucinations and toward reliable, self‑aware reasoning.
If you mean pure as in there’s not additional training beyond the pretraining, I don’t think any model has been pure since gpt-3.5.
This seems to be one of the brutal truths of the modern world, and as far as I can tell it applies to everything. There's always a race to the bottom to make everything as cheaply as possible, and the further the industry goes down that "cheapness" scale, the more "quality" loses market share, the more expensive "quality" must be in order to operate at all, and finally things that used to be just "normal" and not too expensive are now luxury goods.
Consider textiles, carpentry, masonry, machine tooling, appliances, etc. etc.
This doesn't feel like a good outcome, but I'm not sure there's anything that can be done about it.
Instead of broad employment of artisan breadsmiths, we have people doing email work, because it’s more economically valuable. If the government mandated a higher quality of bread, we’d be slightly richer and bread and slightly poorer in everything else.
He was extremely kind, gave me a lot of interesting life advice. I remember him saying that he got most of his ideas just from playing around with mechanics and experimenting a lot, he was never really one to get grand visions.
Anyways, great fellow, glad he opened source V (as he called it).
Dead Comment
I listened to Lex Friedman for a long time, and there was a lot of critiques of him (Lex) as an interviewer, but since the guests were amazing, I never really cared.
But after listening to Dwarkesh, my eyes are opened (or maybe my soul). It doesn't matter I've heard of not-many of his guests, because he knows exactly the right questions to ask. He seems to have genuine curiosity for what the guest is saying, and will push back if something doesn't make sense to him. Very much recommend.
I am curious what the user interfaces of AI in the future will be, I think whoever can crack that will create immense value.