Readit News logoReadit News
arrakark commented on The Enterprise Experience   churchofturing.github.io/... · Posted by u/Improvement
arrakark · 12 days ago
Love it. Describes my new job at $ENTERPRISE very well.
arrakark commented on Test Results for AMD Zen 5   agner.org/forum/viewtopic... · Posted by u/matt_d
ashvardanian · a month ago
> All vector units have full 512 bits capabilities except for memory writes. A 512-bit vector write instruction is executed as two 256-bit writes.

That sounds like a weird design choice. Curious if this will affect memcpy-heavy workloads.

Writes aside, Zen5 is taking much longer to roll out than I thought, and some of AMD's positioning is (almost expectedly) misleading, especially around AI.

AMD's website claims Zen5 is the "Leading CPU for AI" (<https://www.amd.com/en/products/processors/server/epyc/ai.ht...>), but I strongly doubt that. First, they compare Zen5 (9965), which is still largely unavailable, to Xeon2 (8280), a 2 generations older processor. Xeon4 is abundantly available and comes with AMX, an exclusive feature to Intel. I doubt AVX-512 support with a 512-bit physical path and even twice as many cores will be enough to compete with that (if we consider just the ALU throughput rather than the overall system & memory).

arrakark · a month ago
Cache-line bursts/beats tend to be standardized to 64B in lots of NoC architectures.
arrakark commented on Side-Effect of Trump Flip-Flopping on Tariffs   pomogaev.ca/trump_tarrif/... · Posted by u/arrakark
arrakark · 5 months ago
Mods please delete if too political. I just thought it's an interesting idea.
arrakark commented on NoProp: Training neural networks without back-propagation or forward-propagation   arxiv.org/abs/2503.24322... · Posted by u/belleville
itsthecourier · 5 months ago
"Whenever these kind of papers come out I skim it looking for where they actually do backprop.

Check the pseudo code of their algorithms.

"Update using gradient based optimizations""

arrakark · 5 months ago
Same.

If I had to guess it's just local gradients, not an end-to-end gradient.

arrakark commented on Aptera's First Solar Road Trip. 300 Miles, One Charge   aptera.us/apteras-first-s... · Posted by u/geox
arrakark · 5 months ago
I think Aptera's vision is fantastic. People are quick to judge the looks or three-wheelness of the car, but I think that's just a symptom of cars being marketed as status symbols nowadays.

I am worried about production. With all of the years it's taking them to get there, they can run out of money at any time it seems. It's unknown if they can raise enough money on their terms to get this thing to production.

arrakark commented on A bear case: My predictions regarding AI progress   lesswrong.com/posts/oKAFF... · Posted by u/suryao
gymbeaux · 6 months ago
Yeah agree 100%. LLMs are overrated. I describe them as the “Jack of all, master of none” of AI. LLMs are that jackass guy we all know who has to chime in to every topic like he knows everything, but in reality he’s a fraud with low self-esteem.

I’ve known a guy since college who now has a PhD in something niche, supposedly pulls a $200k/yr salary. One of our first conversations (in college, circa 2014) was how he had this clever and easy way to mint money- by selling Minecraft servers installed on Raspberry Pis. Some of you will recognize how asinine this idea was and is. For everyone else- back then, Minecraft only ran on x86 CPUs (and I doubt a Pi would make a good Minecraft server today, even if it were economical). He had no idea what he was talking about, he was just spewing shit like he was God’s gift. Actually, the problem wasn’t that he had no idea- it was that he knew a tiny bit- enough to sound smart to an idiot (remind you of anyone?).

That’s an LLM. A jackass with access to Google.

I’ve had great success with SLMs (small language models), and what’s more I don’t need a rack of NVIDIA L40 GPUs to train and use them.

arrakark · 6 months ago
But original MC ran on JVM, which can run on ARM...
arrakark commented on A bear case: My predictions regarding AI progress   lesswrong.com/posts/oKAFF... · Posted by u/suryao
hangonhn · 6 months ago
I'll add this in case it's helpful to anyone else: LLMs are really good at regex and undoing various encodings/escaping, especially nested ones. I would go so far to say that it's better than a human at the latter.

I once spend over an hour trying to unescape JSON containing UTF8 values that's been escaped prior to being written to AWS's Cloudwatch Logs for MySQL audit logs. It was a horrific level of pain until I just asked ChatGPT to do it and it figured out all the series of escapes and encoding immediately and gave me the step to reverse them all.

LLM as a sidekick has saved me so much time. I don't really use it to generate code but for some odd tasks or API look up, it's a huge time saver.

arrakark · 6 months ago
> LLMs are really good at regex

Maybe that's changed recently, but I have struggled to get all but the most basic regex working from GPT-4o-mini

u/arrakark

KarmaCake day27June 1, 2023View Original