Readit News logoReadit News
imatrix commented on LLMs tell bad jokes because they avoid surprises   danfabulich.medium.com/ll... · Posted by u/dfabulich
imatrix · 4 months ago
it's not about the punchline, it's about the fact that jokes are an extremely complicated thing. The reason? It's a highly abstract structure of information, not unlike low-level speech (how words are supposed to be put together in different languages), code (different programming languages have different rules for commands and sequences).

Even the best LLm's are still at the level of children in this abstraction, so they can't make "quality" jokes. They also suffer from not having a unique personality and being the average of everything. Until this is addressed, don't expect great jokes to come out of AI. It's almost the most challenging discipline. I wouldn't be afraid to make the comparison that the "joke" is the real Turing test.

imatrix commented on Show HN: WebGPU enables local LLM in the browser – demo site with AI chat   andreinwald.github.io/bro... · Posted by u/andreinwald
imatrix · 5 months ago
Its not working.

Loading model: waiting WebGPU error: WebGPU is NOT supported on this browser.

Im on macOS and Safari.

imatrix · 5 months ago
Same on iOS and mobile Safari - not working.
imatrix commented on Show HN: WebGPU enables local LLM in the browser – demo site with AI chat   andreinwald.github.io/bro... · Posted by u/andreinwald
imatrix · 5 months ago
Its not working.

Loading model: waiting WebGPU error: WebGPU is NOT supported on this browser.

Im on macOS and Safari.

imatrix commented on Apple M3 Ultra   apple.com/newsroom/2025/0... · Posted by u/ksec
adgjlsfhk1 · 10 months ago
This chip has 0GB vram. It has 8 channel lpddr5.
imatrix · 10 months ago
This is not correct. It is like VRAM, not like normal PC RAM.

Lowest latency of DDR5-6400 on normal PC starting at 60ns+

Lowest latency of VRAM on GeForce RTX 4090 starting at 14 ns

Lowest latency of Apple M1 Memory starting at 5 ns, its more like L3 cache

And on Apple M chip, this ultrafast memory is available for CPU, GPU and NPU.

https://www.anandtech.com/show/17024/apple-m1-max-performanc...https://chipsandcheese.com/p/microbenchmarking-nvidias-rtx-4...

imatrix commented on China's AI Earthquake: How DeepSeek's Surprise Model R1 Shook Silicon Valley   neuronad.com/featured/chi... · Posted by u/imatrix
F7F7F7 · a year ago
So OpenAI doesn’t need to burn through a trillion dollars in the next 10 years after all?

Cool.

imatrix · a year ago
I think that if investors see this, that you only need few millions to be almost number one, they will not give them billions so easily. Deepseek changing the narrative of spending money in AI tech.
imatrix commented on China's AI Earthquake: How DeepSeek's Surprise Model R1 Shook Silicon Valley   neuronad.com/featured/chi... · Posted by u/imatrix
bobx11 · a year ago
> The Chinese lab reportedly needed only two months of training time to reach performance levels that took Google, Microsoft, Meta, and OpenAI years to develop.

Sounds a lot like standing on the backs of giants to me. Why would this blow minds that with newer compute and full hindsight, someone could reproduce something more efficiently?

I feel like I’m missing the point and Google didn’t illuminate any deep article that represented this achievement in novel terms.

imatrix · a year ago
Right now, they are standing on the backs of giants, but in the future...it could be opposite.
imatrix commented on China's AI Earthquake: How DeepSeek's Surprise Model R1 Shook Silicon Valley   neuronad.com/featured/chi... · Posted by u/imatrix
imatrix · a year ago
What they achieved with so few resources, it’s WOW and WTF. Maybe it will cause end of the open source models? Because it’s always about money in the end.

u/imatrix

KarmaCake day3June 18, 2014View Original