Readit News logoReadit News
mxwsn commented on Show HN: The Hessian of tall-skinny networks is easy to invert   github.com/a-rahimi/hessi... · Posted by u/rahimiali
tubs · a month ago
You explain well so what I never understood is how the Jacobians aren't the first derivatives themselves?

Also if you have happen to have any suggestions for linear algebra for someone who uses it without really understanding it (I can write a measurement function for an EKF from scratch OK, but I don't really understand why the maths does what it does) I would really appreciate it.

mxwsn · a month ago
The Jacobian is first derivatives, but for a function mapping N to M dimensions. It's the first derivative of every output wrt every input, so it will be an N x M matrix.

The gradient is a special case of the Jacobian for functions mapping N to 1 dimension, such as loss functions. The gradient is an N x 1 vector.

mxwsn commented on How does gradient descent work?   centralflows.github.io/pa... · Posted by u/jxmorris12
mxwsn · 4 months ago
Wow! The title suggests introductory material, but in my opinion this has strong potential to win test of time awards for research.
mxwsn commented on Sora 2   openai.com/index/sora-2/... · Posted by u/skilled
minimaxir · 4 months ago
This Sora 2 generation of Cyberpunk 2077 gameplay managed to reproduce it extremely closely, which is baffling: https://x.com/elder_plinius/status/1973124528680345871

> How the FUCK does Sora 2 have such a perfect memory of this Cyberpunk side mission that it knows the map location, biome/terrain, vehicle design, voices, and even the name of the gang you're fighting for, all without being prompted for any of those specifics??

> Sora basically got two details wrong, which is that the Basilisk tank doesn't have wheels (it hovers) and Panam is inside the tank rather than on the turret. I suppose there's a fair amount of video tutorials for this mission scattered around the internet, but still––it's a SIDE mission!

Everyone already assumed that Sora was trained on YouTube, but "generate gameplay of Cyberpunk 2077 with the Basilisk Tank and Panam" would have generated incoherent slop in most other image/video models, not verbatim gameplay footage that is consistent.

For reference, this is what you get when you give the same prompt to Veo 3 Fast (trained by the company that owns YouTube): https://x.com/minimaxir/status/1973192357559542169

mxwsn · 4 months ago
That's really interesting. What if they RAG search related videos from the prompt, and condition on that to generate? That might explain fidelity like this
mxwsn commented on Diffusion Beats Autoregressive in Data-Constrained Settings   blog.ml.cmu.edu/2025/09/2... · Posted by u/djoldman
tripplyons · 5 months ago
At that point it is not following a diffusion training objective. I am aware of papers that do this, but I have not seen one that shows it as a better pretraining objective than something like v-prediction or flow matching.
mxwsn · 5 months ago
Why is not the diffusion training objective? The technique is known as self-conditioning right? Is it an issue with conditional Tweedie's?
mxwsn commented on AI is different   antirez.com/news/155... · Posted by u/grep_it
mxwsn · 6 months ago
AI with ability but without responsibility is not enough for dramatic socioeconomic change, I think. For now, the critical unique power of human workers is that you can hold them responsible for things.

edit: ability without accountability is the catchier motto :)

mxwsn commented on Unlike ChatGPT, Anthropic has doubled down on Artifacts   ben-mini.com/2025/claude-... · Posted by u/bewal416
mxwsn · 7 months ago
Has anyone come across any really cool artifacts? I'd be curious to see
mxwsn commented on Web3 Onboarding Was a Flop – and Thank Goodness   tomhadley.link/blog/web3-... · Posted by u/solumos
gregmac · 7 months ago
Am I the only one struggling to decipher this?

I thought web3 was supposed to be some kind of decentralized compute, where rather than run on your own hardware or IaaS/PaaS you could make use of compute resources that vary wildly day-to-day in availability, performance, and cost, because they were somehow also mining rigs or something? But it's "decentralized" because there's not one entity running the thing.

There is not a mention of that in the article.

Is it actually supposed to just be microtranscations paid with cryptocurrency? Where's the "decentralized" part of that?

Anyway, instead the best I can see this article seems to be talking about how it turns out people aren't using blockchain for buying things, and makes the (apparently) shocking conclusion "the one thing people always wanted: money that just works."

mxwsn · 7 months ago
Stablecoins transferred $27 trillion in 2024 - more than Visa and Mastercard combined. This is right in the article.

Stablecoins operate using decentralized ledgers on e.g. Ethereum which use decentralized compute. This isn't mentioned explicitly because the target audience knows this already.

mxwsn commented on Claude 4   anthropic.com/news/claude... · Posted by u/meetpateltech
throwaway314155 · 9 months ago
Gemini can beat the game?
mxwsn · 9 months ago
Gemini has beat it already, but using a different and notably more helpful harness. The creator has said they think harness design is the most important factor right now, and that the results don't mean much for comparing Claude to Gemini.
mxwsn commented on The booming, high-stakes arms race of airline safety videos   thehustle.co/originals/th... · Posted by u/gmays
mxwsn · 10 months ago
Huh, I imagined this was because of relaxing regulation.
mxwsn commented on Deep Learning Is Not So Mysterious or Different   arxiv.org/abs/2503.02113... · Posted by u/wuubuu
cgdl · a year ago
Agreed, but PAC-Bayes or other descendants of VC theory is probably not the best explanation. The notion of algorithmic stability provides a (much) more compelling explanation. See [1] (particularly Sections 11 and 12)

[1] https://arxiv.org/abs/2203.10036

mxwsn · a year ago
Good read, thanks for sharing

u/mxwsn

KarmaCake day1616September 15, 2017View Original