Readit News logoReadit News
gojomo commented on Study mode   openai.com/index/chatgpt-... · Posted by u/meetpateltech
no_wizard · a month ago
>Now, everyone basically has a personal TA, ready to go at all hours of the day

This simply hasn't been my experience.

Its too shallow. The deeper I go, the less it seems to be useful. This happens quick for me.

Also, god forbid you're researching a complex and possibly controversial subject and you want it to find reputable sources or particularly academic ones.

gojomo · a month ago
Grandparent testimony of success, & parent testimony of frustration, are both just wispy random gossip when they don't specify which LLMs delivered the reported experiences.

The quality varies wildly across models & versions.

With humans, the statement "my tutor was great" and "my tutor was awful" reflect very little on "tutoring" in general, and are barely even responses to each other withou more specificity about the quality of tutor involved.

Same with AI models.

gojomo commented on GLM-4.5: Reasoning, Coding, and Agentic Abililties   z.ai/blog/glm-4.5... · Posted by u/GaggiX
ed · a month ago
Subliminal learning happens when the teacher and student models share a common base model, which is unlikely to be the case here
gojomo · a month ago
Given that other work shows that models often converge on similar internal representations, I'd not be surprised if there were close analogues of 'subliminal learning' that don't require shared-ancestor-base-model, just enough overlap in training material.

Further, "enough" training from another model's outputs – de facto 'distillation' – is likely to have similar effects as starting from a common base model, just "from thge other direction".

(Finally: some of the more nationalistic-paranoid observers seem to think Chinese labs have relied on exfiltrated weights from US entities. I don't personally think that'd be a likely or necessary contributor to Z.ai & others' successes, the mere appearance of this occasional "I am Claude" answer is sure to fuel further armchair belief in those theories.)

gojomo commented on Ask HN: Is it time to fork HN into AI/LLM and "Everything else/other?"    · Posted by u/bookofjoe
azath92 · 2 months ago
Its pretty easy to do the user-loyal bit, with a bit of prompting to give an llm your preferences/profile. Not ideologically loyal, but i mean acting in accordance with your interests.

The tricky part is having that act across all sites in a light and seamless way. Ive been working on a HN reskin, and it only is fast/transparent/cheap enough because HN has an api (no scraping needed), and the titles are descriptive enough that you can filter based on them, as simonws demo does. But its still HN specific.

I dont know if llms are fast enough at the moment to do this on the fly for arbitrary sites, but steps in that direction are interesting!

gojomo · 2 months ago
I'd expect a noticeable delay with current local LLMs - especially visiting a site for the 1st time. But then they could potentially memoize their heuristics for certain designs, including recognzing when some "deeper thought" newly required by server-side redesigns.

But of course local GPU processing power, & optimizations for LLM-like tools, all adancing rapidly. And these local agents could potentially even outsource tough decisions to heavierweight remote services. Essentially, they'd maintain/reauthor your "custom extension", themselves using other models, as necessary.

And forward-thinking sites might try to make that process easier, with special APIs/docs/recipe-interchanges for all users' agents to share their progress on popular needs.

gojomo commented on Ask HN: Is it time to fork HN into AI/LLM and "Everything else/other?"    · Posted by u/bookofjoe
gojomo · 2 months ago
~simonw's demo of a quickie customized HN front-end is great.

But ultimately, your browser should have a local, open-source, user-loyal LLM that's able to accept human-language descriptions of how you'd like your view of some or all sites to change, and just like old Greasemonkey scripts or special-purpose extensions, it'd just do it, in the DOM.

Then instead of needing to raise this issue via an "Ask HN", you'd just tell your browser: "when I visit HN, hide all the AI/LLM posts".

gojomo commented on Measuring the impact of AI on experienced open-source developer productivity   metr.org/blog/2025-07-10-... · Posted by u/dheerajvs
narush · 2 months ago
You can see this analysis in the factor analysis of "Below-average use of AI tools" (C.2.7) in the paper [1], which we mark as an unclear effect.

TLDR: over the first 8 issues, developers do not appear to get majorly less slowed down.

[1] https://metr.org/Early_2025_AI_Experienced_OS_Devs_Study.pdf

gojomo · 2 months ago
Thanks, that's great!

But: if all developers did 136 AI-assisted issues, why only analyze excluding the 1st 8, rather than, say, the first 68 (half)?

gojomo commented on Measuring the impact of AI on experienced open-source developer productivity   metr.org/blog/2025-07-10-... · Posted by u/dheerajvs
narush · 2 months ago
Hey Simon -- thanks for the detailed read of the paper - I'm a big fan of your OS projects!

Noting a few important points here:

1. Some prior studies that find speedup do so with developers that have similar (or less!) experience with the tools they use. In other words, the "steep learning curve" theory doesn't differentially explain our results vs. other results.

2. Prior to the study, 90+% of developers had reasonable experience prompting LLMs. Before we found slowdown, this was the only concern that most external reviewers had about experience was about prompting -- as prompting was considered the primary skill. In general, the standard wisdom was/is Cursor is very easy to pick up if you're used to VSCode, which most developers used prior to the study.

3. Imagine all these developers had a TON of AI experience. One thing this might do is make them worse programmers when not using AI (relatable, at least for me), which in turn would raise the speedup we find (but not because AI was better, but just because with AI is much worse). In other words, we're sorta in between a rock and a hard place here -- it's just plain hard to figure out what the right baseline should be!

4. We shared information on developer prior experience with expert forecasters. Even with this information, forecasters were still dramatically over-optimistic about speedup.

5. As you say, it's totally possible that there is a long-tail of skills to using these tools -- things you only pick up and realize after hundreds of hours of usage. Our study doesn't really speak to this. I'd be excited for future literature to explore this more.

In general, these results being surprising makes it easy to read the paper, find one factor that resonates, and conclude "ah, this one factor probably just explains slowdown." My guess: there is no one factor -- there's a bunch of factors that contribute to this result -- at least 5 seem likely, and at least 9 we can't rule out (see the factors table on page 11).

I'll also note that one really important takeaway -- that developer self-reports after using AI are overoptimistic to the point of being on the wrong side of speedup/slowdown -- isn't a function of which tool they use. The need for robust, on-the-ground measurements to accurately judge productivity gains is a key takeaway here for me!

(You can see a lot more detail in section C.2.7 of the paper ("Below-average use of AI tools") -- where we explore the points here in more detail.)

gojomo · 2 months ago
Did each developer do a large enough mix of AI/non-AI tasks, in varying orders, that you have any hints in your data whether the "AI penalty" grew or shrunk over time?
gojomo commented on Public Signal Backups Testing   community.signalusers.org... · Posted by u/blendergeek
Nextgrid · 2 months ago
This uses the same protocol as macOS, so same outcome. However (not sure if it helps in this case), keep in mind that different data is retained depending on whether the backup is encrypted or not (the tool supports encrypted backups too). So maybe check with encryption enabled and see if it carries over Signal messages?
gojomo · 2 months ago
It doesn't. Signal opts its data out of even encrypted local iPhone backups.
gojomo commented on Public Signal Backups Testing   community.signalusers.org... · Posted by u/blendergeek
gojomo · 2 months ago
News to me! Where is this option described, ideally by Signal itself?

If you are alleging that Apple's own local Finder/Itunes backup of an iPhone includes Signal messages, that's not true, against reasonable user expectations, by Signal's own design choices.

Anyone who's counting on such local backups to save their histories is in for the same rude surprise I and many others have hit unaware:

https://www.reddit.com/r/signal/comments/1hgukpg/backup_and_...

gojomo · 2 months ago
And note: Signal has been misleading users about options for iOS backups for a decade! In September 2015, users were told it was "on the roadmap": https://github.com/signalapp/Signal-iOS/issues/905#issuecomm...

No option has ever existed on iOS, despite the recent announcement assuring users "Local backups still exist". They've never existed for iOS users!

gojomo commented on Public Signal Backups Testing   community.signalusers.org... · Posted by u/blendergeek
Nextgrid · 2 months ago
"libimobiledevice" has CLI tools that can backup/restore iOS devices on Linux.
gojomo · 2 months ago
Do such backups retain Signal message history - unlike Apple's native local backups to MacOS?

u/gojomo

KarmaCake day30026February 24, 2007
About
Gordon Mohr, maker of software.

Twitter: http://twitter.com/gojomo

Project: http://thunkpedia.org

Idea blog: http://memesteading.com

Older blog: http://gojomo.blogspot.com

Homepage: http://xavvy.com (username @ here for email contact)

My HN peeve is formulaic downbeat comments, like: "How is this news, I already knew this!" "…Betteridge's Law…" "I stopped reading at…"

View Original