If you're talking about longevity of solar or wind turbines, given that their lifespan is measured in tens of years, it does give you some ability to respond to some external group cutting your supply off. It's not remotely the same as oil/gas dependency.
It’s important in the fair use assessment to understand that the training itself is fair use, but the pirating of the books is the issue at hand here, and is what Anthropic “whoopsied” into in acquiring the training data.
Buying used copies of books, scanning them, and training on it is fine.
Rainbows End was prescient in many ways.
I think that we devs are now very skeptical because we are the ones that have to fix the sh that llms spit out. But likely we will be taken out of the loop completely.
LLMs are hardly reliable ways to reproduce copyrighted works. The closest examples usually involve prompting the LLM with a significant portion of the copyrighted work and then seeing it can predict a number of tokens that follow. It’s a big stretch to say that they’re reliably reproducing copyrighted works any more than, say, a Google search producing a short excerpt of a document in the search results or a blog writer quoting a section of a book.
It’s also interesting to see the sudden anti-LLM takes that twist themselves into arguing against tools or platforms that might reproduce some copyrighted content. By this argument, should BitTorrent also be banned? If someone posts a section of copyrighted content to Hacker News as a comment, should YCombinator be held responsible?
If you cannot see the difference between BitTorrent and Ai models, then it's probably not worth engaging with you.
But Ai model have been shown to reproduce the training data
https://gizmodo.com/ai-art-generators-ai-copyright-stable-di...
One of the key aspects of the act is how a model provider is responsible if the downstream partners misuse it in any way. For open source, it's a very hard requirement[1].
> GPAI model providers need to establish reasonable copyright measures to mitigate the risk that a downstream system or application into which a model is integrated generates copyright-infringing outputs, including through avoiding overfitting of their GPAI model. Where a GPAI model is provided to another entity, providers are encouraged to make the conclusion or validity of the contractual provision of the model dependent upon a promise of that entity to take appropriate measures to avoid the repeated generation of output that is identical or recognisably similar to protected works.
[1] https://www.lw.com/en/insights/2024/11/european-commission-r...
It was a decade too late and written by people who were incredibly out of touch with the actual problem. The GDPR is a bit better, but it's still a far bigger nuisance for regular European citizens than the companies that still largely unhindered track and profile the same.
The economic cases for manned moon or Mars programs look really iffy these days. The US has poured tens of $billions down the SLS rat hole, with very little to show for it. And Wienersmith's A City on Mars is a pretty damning dissection of the whole concept of Martian colonies.
It's pretty difficult to predict what spinoffs would come from attempting to put a colony on Mars. I would imagine to succeed we would need to solve a lot of challenges with human biology, genetic engineering, automation, and many novel engineering solutions.
But economics is not the only reason to do things, and I bet you don't expect everything humans do to have a purely economic rational.