Readit News logoReadit News
FuckButtons commented on Google is watching   not-a-tech-bro.ghost.io/g... · Posted by u/hodgesrm
brettgriffin · 8 days ago
What I find interesting about this article is that the author is the founder of Cloudera. Cloudera was one of the pioneers of the Hadoop ecosystem (and de facto data lake movements).

For the uninitiated, data lakes are used to centralize vast quantities of data - often consumer data - usually by large organizations and governments to provide insights and inform decisions within the organization. Some call this surveillance capitalism.

Cloudera IPO'd somewhere around $2B and was taken private in a deal led by KKR for around $5B.

FuckButtons · 8 days ago
Other than possibly loosing social capital, why does this surprise you? they are intimately familiar with the intricacies of how the technology is used to hoard and catalogue every aspect of our digital lives.
FuckButtons commented on MCP doesn't need tools, it needs code   lucumr.pocoo.org/2025/8/1... · Posted by u/the_mitsuhiko
anon7000 · 9 days ago
Exactly. LLMs are trained on huge amounts of bash scripts. They “know” how to use grep/awk/whatever. ASTs are, I assume, not really part of that training data. How would they know how to work well with on? LLMs are trained on what humans do to code. Yes, I assume down the road someone will train more efficient versions that can work more closely with the machine. But LLMs work as well as they do because they have a large body of “sed” statements in their statistical models
FuckButtons · 8 days ago
treesitter is more or less a universal AST parser you can run queries against. Writing queries against an AST that you incrementally rebuild is massively more powerful and precise in generating the correct context than manually writing infinitely many shell pipeline oneliners and correctly handling all of the edge cases.
FuckButtons commented on The lottery ticket hypothesis: why neural networks work   nearlyright.com/how-ai-re... · Posted by u/076ae80a-3c97-4
xg15 · 8 days ago
Wouldn't this imply that most of the inference time storage and compute might be unnecessary?

If the hypothesis is true, it makes sense to scale up models as much as possible during training - but once the model is sufficiently trained for the task, wouldn't 99% of the weights be literal "dead weight" - because they represent the "failed lottery tickets", i.e. the subnetworks that did not have the right starting values to learn anything useful? So why do we keep them around and waste enormous amounts of storage and compute on them?

FuckButtons · 8 days ago
For any particular single pattern learned 99% of the weights are dead weight. But it’s not the same 99% for each lesson learned.
FuckButtons commented on Swarm robotics could spell the end of the assembly line   therobotreport.com/swarm-... · Posted by u/speckx
FuckButtons · 22 days ago
Someone’s been playing factorio.
FuckButtons commented on Problems the AI industry is not addressing adequately   thealgorithmicbridge.com/... · Posted by u/baylearn
visarga · 2 months ago
It's not about achieving AGI as a final product, it's about building a perpetual learning machine fueled by real-time human interaction. I call it the human-AI experience flywheel.

People bring problems to the LLM, the LLM produces some text, people use it and later return to iterate. This iteration functions as a feedback for earlier responses from the LLM. If you judge an AI response by the next 20 rounds of interaction or more you can gauge if it was useful or not. They can create RLHF data this way, using hindsight or extra context from other related conversations of the same user on the same topic. That works because users try the LLM ideas in reality and bring outcome results back to the model, or they simply recall from their personal experience if that approach would work or not. The system isn't just built to be right; it's built to be correctable by the user base, at scale.

OpenAI has 500M users, if they generate 1000 tokens/user/day that means 0.5T interactive tokens/day. The chat logs dwarf the original training set in size and are very diverse, targeted to our interests, and mixed with feedback. They are also "on policy" for the LLM, meaning they contain corrections to mistakes the LLM made, not generic information like web scrape.

You're right that LLMs eventually might not even need to crawl the web, they have the whole society dump data into their open mouths. That did not happen with web search engines, only social networks did that in the past. But social networks are filled with our cultural wars and self conscious posing, while the chat room is an environment where we don't need to signal our group alignment.

Web scraping gives you humanity's external productions - what we chose to publish. But conversational logs capture our thinking process, our mistakes, our iterative refinements. Google learned what we wanted to find, but LLMs learn how we think through problems.

FuckButtons · 2 months ago
I see where you’re coming from, but I think teasing out something that looks like a clear objective function that generalizes to improved intelligence from llm interaction logs is going to be hellishly difficult. Consider, that most of the best llm pre training comes from being very very judicious with the training data, selecting the right corpus of llm interaction logs and then defining an objective function that correctly models…? Being helpful? From that sounds far harder than just working from scratch with rlhf.
FuckButtons commented on Everything around LLMs is still magical and wishful thinking   dmitriid.com/everything-a... · Posted by u/troupo
ls612 · 2 months ago
Even when quantized down to 4 bits to fit on a 4090?
FuckButtons · 2 months ago
Not in my experience, running qwen3:32b is good, but it’s not as coherent or useful as 3.5 at a 4bit quant. But the gap is a lot narrower than llama 70b.
FuckButtons commented on LLM code generation may lead to an erosion of trust   jaysthoughts.com/aithough... · Posted by u/CoffeeOnWrite
HardCodedBias · 2 months ago
All of this fighting against LLMs is pissing in the wind.

It seems that LLMs, as they work today, make developers more productive. It is possible that they benefit less experienced developers even more than experienced developers.

More productivity, and perhaps very large multiples of productivity, will not be abandoned due roadblocks constructed by those who oppose the technology due to some reason.

Examples of the new productivity tool causing enormous harm (eg: bug that brings down some large service for a considerable amount of time) will not stop the technology if it being considerable productivity.

Working with the technology and mitigating it's weaknesses is the only rational path forward. And those mitigation can't be a set of rules that completely strip the new technology of it's productivity gains. The mitigations have to work with the technology to increase its adoption or they will be worked around.

FuckButtons · 2 months ago
I think you’re arguing against something the author didn’t actually say.

You seem to be claiming that this is a binary, either we will or won’t use llms, but the author is mostly talking about risk mitigation.

By analogy it seems like you’re saying the author is fundamentally against the development of the motor car because they’ve pointed out that some have exploded whereas before, we had horses which didn’t explode, and maybe we should work on making them explode less before we fire up the glue factories.

FuckButtons commented on U.S. bombs Iranian nuclear sites   bbc.co.uk/news/live/ckg3r... · Posted by u/mattcollins
barbazoo · 2 months ago
Which stock do I buy
FuckButtons · 2 months ago
Not TSMC.
FuckButtons commented on U.S. bombs Iranian nuclear sites   bbc.co.uk/news/live/ckg3r... · Posted by u/mattcollins
MangoToupe · 2 months ago
So much for humanity learning from its mistakes....
FuckButtons · 2 months ago
Trump doesn’t seem like the kind of person to learn from his, or anyone else's mistakes.
FuckButtons commented on U.S. bombs Iranian nuclear sites   bbc.co.uk/news/live/ckg3r... · Posted by u/mattcollins
dj_gitmo · 2 months ago
It’s horrible that the president can start a war without even asking congress.
FuckButtons · 2 months ago
The strong do as they will while the weak suffer what they must.

I’m glad that trump has returned us to a world where quotes from the 5th century bc seem like commentary on current affairs, since it means that all my time learning about power dynamics in political systems during antiquity is now completely relevant to dealing with current events, rather than a giant waste of time.

u/FuckButtons

KarmaCake day806July 22, 2019View Original