I was working in this space! And I got fired for refusing to work on more upsell features for clients like Coca Cola and such.
I don't want to work on adding fucking ADS into checkout. That is fucked up.
I was working in this space! And I got fired for refusing to work on more upsell features for clients like Coca Cola and such.
I don't want to work on adding fucking ADS into checkout. That is fucked up.
One thing I’ve thought about is how AI assistants are actually turning code into literature, and literature into code.
In old-fashioned programming, you can roughly observe a correlation between programmer skill and linear composition of their programs, as in, writing it all out at once from top to bottom without breaks. There was then this pre-modern era where that practice was criticized in favor of things like TDD and doc-first and interfaces, but it still probably holds on the subtasks of those methods. Now there are LLM agents that basically operate the same way. A stronger model will write all at once, while a weaker model will have to be guided through many stages of refinement. Also, it turns the programmer into a literary agent, giving prose descriptions piece by piece to match the capabilities of the model, but still in linear fashion.
And I can’t help but think that this points to an inadequacy of the language. There should be a programming language that enables arbitrary complexity through deterministic linear code, as humans seem to have an innate comfort with. One question I have about this is why postfix notation is so unpopular versus infix or prefix, where complex expressions in postfix read more like literature where details build up to greater concepts. Is it just because of school? Could postfix fix the stem/humanities gap?
I see LLMs as translators, which is not new because that’s what they were built for, but in this case between two very different structures of language, which is why they must grow in parameters with the size of the task rather than process linearly along a task with limited memory, as in the original spoken language to spoken language task. If mathematics and programming were more like spoken language, it seems the task would be massively simpler. So maybe the problem for us too is the language and not the intelligence.
I'm curious what the Google Docs integration is. IDEs should be a lot more like Google Docs: cloud first, continuous save, multi-player, commenting, total revision history, etc. I would love to write working script code in Google Docs and instantly have access to it via a url.
It's made me very lazy with my thinking and writing.
He gives a strongly NVidia oriented answer that I happen to think is dead wrong. Pushing more and more GPU/Memory bandwidth into more and more expensive packages that are obsolete after a year or two isn't the approach that I think will win in the end.
I think systems which eliminate the memory/compute distinction completely, like FPGA but more optimized for throughput, instead of latency, are the way to go.
Imagine if you had a network of machines, that could each handle one layer of an LLM with no memory transfers, your bottleneck would be just getting the data between layers. GPT 4, for example, is likely a 8 separate columns of 120 layers of of 1024^2 parameter matrix multiplies. Assuming infinitely fast compute, you still have to transfer at least 2KB of parameters between layers for every token. Assuming PCI Express 7, at about 200 Gigabytes/second, that's about 100,000,000 tokens/second across all of the computing fabric.
Flowing 13 trillion tokens through that would take 36 hours/epoch.
Doing all of that in one place is impressive. But if you can farm it out, and have a bunch of CPUs and network connections, you're transferring 4k each way for each token from each workstation. It wouldn't be unreasonable to aggregate all of those flows across the internet without the need for anything super fancy. Even if it took a month/epoch, it could keep going for a very long time.
The simplest solution is to wait until the cost of hashing exceeds the value of your transaction by some reasonable factor. I expect that better solutions will come along by soft fork without adverse effect on supply or decentralization.
2. What are they doing? AGI/ASI is a neat trick, but then what? I’m not asking because I don’t think there is an answer; I’m asking because I want the REAL answer. Larry Ellison was talking about RNA cancer vaccines. Well, I was the one that made the neural network model for the company with the US patent on this technique, and that pitch makes little sense. As the problem is understood today, the computational problems are 99% solved with laptop-class hardware. There are some remaining problems that are not solved by neural networks, but by molecular dynamics, which are done in FP64. Even if FP8 neural structure approximation speeds it up 100x, FP64 will be 99% of the computation. So what we today call “AI infrastructure” is not appropriate for the task they talk about. What is it appropriate for? Well, I know that Sam is a bit uncreative, so I assume he’s just going to keep following the “HER” timeline and make a massive playground for LLMs to talk to each other and leave humanity behind. I don’t think that is necessarily unworthy of our Apollo-scale commitment, but there are serious questions about the honest of the project, and what we should demand for transparency. We’re obviously headed toward a symbiotic merger where LLMs and GenAI are completely in control of our understanding of the world. There is a difference between watching a high-production movie for two hours, and then going back to reality, versus a never-ending stream of false sensory information engineered individually to specifically control your behavior. The only question is whether we will be able to see behind the curtain of the great Oz. That’s what I mean by transparency. Not financial or organizational, but actual code, data, model, and prompt transparency. Is this a fundamental right worth fighting for?