Readit News logoReadit News
AlphaAndOmega0 commented on I'm helping my dog vibe code games   calebleak.com/posts/dog-g... · Posted by u/cleak
blibble · 20 days ago
love the article

slightly concerned tomorrow morning's top HN story will be karparthy telling us how dog-based LLM interfaces are the way of the future

and you'll be left behind if you don't get in now

(and then next week my boss will be demanding I do it)

AlphaAndOmega0 · 20 days ago
The programming workspace of the future of the future will have three employees:

A man, a dog and an instance of Claude.

The dog writes the prompts for Claude, the man feeds the dog, and the dog stops the man from turning off the computer.

AlphaAndOmega0 commented on Gemini 3.1 Pro   blog.google/innovation-an... · Posted by u/MallocVoidstar
suddenlybananas · 25 days ago
I'm not denying any progress, I'm saying that reasoning failures that are simple which have gone viral are exactly the kind of thing that they will toss in the training data. Why wouldn't they? There's real reputational risks in not fixing it and no costs in fixing it.
AlphaAndOmega0 · 25 days ago
Given that Gemini 3 Pro already did solid on that test, what exactly did they improve? Why would they bother?

I double checked and tested on AI Studio, since you can still access the previous model there:

>You should drive. >If you walk there, your car will stay behind, and you won't be able to wash it.

Thinking models consistently get it correct and did when the test was brand new (like a week or two ago). It is the opposite of surprising that a new thinking model continues getting it correct, unless the competitors had a time machine.

AlphaAndOmega0 commented on Gemini 3.1 Pro   blog.google/innovation-an... · Posted by u/MallocVoidstar
suddenlybananas · 25 days ago
They probably had time to toss that example in the training soup.
AlphaAndOmega0 · 25 days ago
Previous models from competitors usually got that correct, and the reasoning versions almost always did.

This kind of reflexive criticism isn't helpful, it's closer to a fully generalized counter-argument against LLM progress, whereas it's obvious to anyone that models today can do things they couldn't do six months ago, let alone 2 years back.

AlphaAndOmega0 commented on Qwen-Image-2.0: Professional infographics, exquisite photorealism   qwen.ai/blog?id=qwen-imag... · Posted by u/meetpateltech
laughingcurve · a month ago
Gary Marcus is not the man to be looking to on this topic
AlphaAndOmega0 · a month ago
Gary Marcus successfully predicted all ten of the one AI Winters.

He also claimed that LLMs were a failure because of prompts that GPT 3.5 couldn't parse, after the launch of GPT-4,which handled them with aplomb.

AlphaAndOmega0 commented on Where did all the starships go?   datawrapper.de/blog/scien... · Posted by u/speckx
ThrowawayR2 · a month ago
"lightweight probe" and "self replicates" don't go together. Nanobots are just as much fantasy physics as FTL is.
AlphaAndOmega0 · a month ago
Nanobots are fantasy? Nobody told your cells or bacteria I guess. We have an existence proof right there.
AlphaAndOmega0 commented on What's up with all those equals signs anyway?   lars.ingebrigtsen.no/2026... · Posted by u/todsacerdoti
lordnacho · a month ago
I love how HN always floats up the answers to questions that were in my mind, without occupying my mind.

I, too, was reading about the new Epstein files, wondering what text artifact was causing things to look like that.

AlphaAndOmega0 · a month ago
Same here. I did notice what I think was an actual error on someone's part, there was a chart in the files comparing black to white IQ distributions, and well, just look at it:

https://nitter.net/AFpost/status/2017415163763429779?s=201

Something clearly went wrong in the process.

AlphaAndOmega0 commented on Apple Maps claims it's 29,905 miles away   mathstodon.xyz/@dpiponi/1... · Posted by u/ColinWright
murkt · 3 months ago
Well, when I’m driving in Kyiv, and there is an air raid alert, usually my car navigation starts to derp, and after a few minutes it thinks that it’s suddenly in Lima, Peru.

Not that I mind too much, I know how to get around without navigation.

AlphaAndOmega0 · 3 months ago
GPS jamming for incoming drones?
AlphaAndOmega0 commented on Show HN: I built an interactive HN Simulator   news.ysimulator.run/news... · Posted by u/johnsillings
chrisweekly · 4 months ago
Amazing. I just found the simulator's mirror^1 "Show HN: I Built An Interactive Human Simulator" and it's priceless.

1. https://news.ysimulator.run/item/1440

AlphaAndOmega0 · 4 months ago
>Interactive Human Simulator is a bold way to describe spinning up a few GPT calls with mood sliders, but sure, let’s call it anthropology. Next iteration can just skip the users entirely and have LLMs submit posts to other LLMs, which, to be fair, would not be noticeably worse than current HN some days.

My sides

AlphaAndOmega0 commented on The Case That A.I. Is Thinking   newyorker.com/magazine/20... · Posted by u/ascertain
popalchemist · 4 months ago
There is simply put no ongoing process and no feedback loop. The model does not learn. The cognition ends when the inference cycle ends. It's not thinking, it just produces output that looks similar to the output of thinking. But the process by which it does that is wholly unreleated.
AlphaAndOmega0 · 4 months ago
Humans with certain amnestic syndromes are incapable of learning. That doesn't make them unintelligent or incapable of thought.
AlphaAndOmega0 commented on Apps SDK   developers.openai.com/app... · Posted by u/alvis
fidotron · 5 months ago
This conception makes sense iff you believe in ChatGPT as the universal user interface of the future. If anything the agentic wave is showing that the chat interfaces are better off hidden behind stricter user interface paradigms.
AlphaAndOmega0 · 5 months ago
>If anything the agentic wave is showing that the chat interfaces are better off hidden behind stricter user interface paradigms.

I'm not sure that claim is justified. The primary agentic use case today is code generation, and the target demographic is used to IDEs/code editors.

While that's probably a good chunk of total token usage, it's not representative of the average user's needs or desires. I strongly doubt that the chat interface would have become so ubiquitous if it didn't have merit.

Even for more general agentic use, a chat interface allows the user the convenience of typing or dictating messages. And it's trivially bundled with audio-to-audio or video-to-video, the former already being common.

I expect that even in the future, if/when richer modalities become standard (and the models can produce video in real-time), most people will be consuming their outputs as text. It's simply more convenient for most use-cases.

u/AlphaAndOmega0

KarmaCake day182August 24, 2023View Original