Readit News logoReadit News
olq_plo commented on Show HN: DenchClaw – Local CRM on Top of OpenClaw   github.com/DenchHQ/DenchC... · Posted by u/kumar_abhirup
olq_plo · 3 days ago
Great, thanks for making me Google what CRM means in this context. Neither your post nor your website explains the acronym.
olq_plo commented on Command-line Tools can be 235x Faster than your Hadoop Cluster (2014)   adamdrake.com/command-lin... · Posted by u/tosh
olq_plo · 2 months ago
And now you can do this with polars in parallel on all your cores and the GPU, using almost the same syntax as in pyspark.
olq_plo commented on Things that aren't doing the thing   strangestloop.io/essays/t... · Posted by u/downboots
olq_plo · 4 months ago
Corollary: Doing the thing and not talking about it in a hammer tweet is also 'not doing the thing'.
olq_plo commented on Parsing webpages with a LLM, revisited   hdembinski.github.io/post... · Posted by u/olq_plo
olq_plo · 5 months ago
I wrote my first post about parsing webpages to structured data with a LLM in January, using local models. Now its October and I did it again with current models and libraries. Boy, what a difference.
olq_plo commented on I almost got hacked by a 'job interview'   blog.daviddodda.com/how-i... · Posted by u/DavidDodda
olq_plo · 5 months ago
The post is so painfully obviously AI written, it hurts my eyes.

The Setup

The Scoop

The Conclusion

I hate AI slop.

olq_plo commented on What Is the Fourier Transform?   quantamagazine.org/what-i... · Posted by u/rbanffy
lutusp · 6 months ago
Such a shame. In an otherwise well-written article, the author mentions Cooley and Tukey's discovery of the FFT, but without mentioning that Gauss discovered it first, among others, each of whom approached the same idea from different directions.

The Wikipedia FFT article (https://en.wikipedia.org/wiki/Fast_Fourier_transform) credits Gauss with originating the FFT idea later expanded on by others, and correctly describes Cooley and Tukey's work as a "rediscovery."

olq_plo · 6 months ago
Yes, bad article to omit that. It is such a cool fun fact. Gauss was unreal.
olq_plo commented on The Bitter Lesson Is Misunderstood   obviouslywrong.substack.c... · Posted by u/JnBrymn
kushalc · 6 months ago
Hey folks, OOP/original author and 20-year HN lurker here — a friend just told me about this and thought I'd chime in.

Reading through the comments, I think there's one key point that might be getting lost: this isn't really about whether scaling is "dead" (it's not), but rather how we continue to scale for language models at the current LM frontier — 4-8h METR tasks.

Someone commented below about verifiable rewards and IMO that's exactly it: if you can find a way to produce verifiable rewards about a target world, you can essentially produce unlimited amounts of data and (likely) scale past the current bottleneck. Then the question becomes, working backwards from the set of interesting 4-8h METR tasks, what worlds can we make verifiable rewards for and how do we scalably make them? [1]

Which is to say, it's not about more data in general, it's about the specific kind of data (or architecture) we need to break a specific bottleneck. For instance, real-world data is indeed verifiable and will be amazing for robotics, etc. but that frontier is further behind: there are some cool labs building foundational robotics models, but they're maybe ~5 years behind LMs today.

[1] There's another path with better design, e.g. CLIP that improves both architecture and data, but let's leave that aside for now.

olq_plo · 6 months ago
Since you seem to know your stuff, why do LLMs need so much data anyway? Humans don't. Why can't we make models aware of their own uncertainty, e.g. feeding the variance of the next token distribution back into the model, as a foundation to guide their own learning. Maybe with that kind of signal, LLMs could develop 'curiosity' and 'rigorousness' and seek out the data that best refines them themselves. Let the AI make and test its own hypotheses, using formal mathematical systems, during training.
olq_plo commented on Tracking Copilot vs. Codex vs. Cursor vs. Devin PR Performance   aavetis.github.io/ai-pr-w... · Posted by u/HiPHInch
rustc · 9 months ago
> “It concludes that the outputs of generative AI can be protected by copyright only where a human author has determined sufficient expressive elements”

How would that work if it's a patch to a project with a copyleft license like GPL which requires all derivate work to be licensed the same?

olq_plo · 9 months ago
IANAL, but it means the commit itself is public domain. When integrated into a code base with a more restrictive license, you can still use that isolated snippet in whatever way you want.

More interesting question is whether one could remove the GPL restrictions on public code by telling AI to rewrite the code from scratch, providing only the behavior of the code.

This could be accomplished by making AI generate a comprehensive test suite first, and then let it write the code of the app seeing only the test suite.

olq_plo commented on TransMLA: Multi-head latent attention is all you need   arxiv.org/abs/2502.07864... · Posted by u/ocean_moist
olq_plo · 10 months ago
Very cool idea. Can't wait for converted models on HF.

u/olq_plo

KarmaCake day10April 14, 2019View Original