Take the C4 training dataset for example. The uncompressed, uncleaned, size of the dataset is ~6TB, and contains an exhaustive English language scrape of the public internet from 2019. The cleaned (still uncompressed) dataset is significantly less than 1TB.
I could go on, but, I think it's already pretty obvious that 1TB is more than enough storage to represent a significant portion of the internet.
That seems implausible.
It really isn't.
What you are likely seeing is HDR which is on most (but not all!) 4K content. The HDR is a separate layer and unrelated to the resolution.
4K versions of films are usually newly restored with modern film scanning - as opposed to the aging masters created for the DVD era that were used to churn out 1st generation Blu-Rays.
The difference between a 4K UHD without HDR and a 1080p Blu-Ray that was recently remastered in 4K from the same source is basically imperceptible from any reasonable viewing distance.
The "visible difference" is mostly better source material, and HDR.
Of course people will convince themselves what they are seeing justifies the cost of the upgrade, just like the $200 audiophile outlet and $350 gold-plated videophile Ethernet cable makes the audio and video really "pop".
People who don't bother to vote for any reason changes elections. It also makes it very hard to make claims about what the majority of Americans want, since so many didn't make their opinions known
A majority of Americans either wanted Trump or didn't care enough to vote against him.
Like this argument just gets absurd: you're claiming building a data center on earth will be harder from a permitting perspective than FAA flight approval for multiple heavy lift rocket launch and landing cycles.
Mining companies routinely open and close enormous surface area mines all over the world and manage permitting for that just fine.
There's plenty of land no one will care if your build anything on, and being remote with maybe poor access roads is still going to be enormously cheaper then launching a state of the art heavy lift rocket which doesn't actually exist yet.
I wonder if this is actually true.
I can't get in detail about real numbers but it's not doable with current hardware by a large margin.
But that’s just me, and I'm not trying to convince anyone.
Honestly I don't know what commenters on hackernews are building, but a few months back I was hoping to use AI to build the interaction layer with Stripe to handle multiple products and delayed cancellations via subscription schedules. Everything is documented, the documentation is a bit scattered across pages, but the information is out there. At the time there was Opus 4.1, so I used that. It wrote 1000 lines of non-functional code with 0 reusability after several prompts. I then asked something to Chat gpt to see if it was possible without using schedules, it told me yes (even if there is not) and when I told Claude to recode it, it started coding random stuff that doesn't exist. I built everything to be functional and reusable myself, in approximately 300 lines of code.
The above is a software engineering problem. Reimplementing a JSON parser using Opus is not fun nor useful, so that should not be used as a metric