Readit News logoReadit News
arkmm commented on Faster asin() was hiding in plain sight   16bpp.net/blog/post/faste... · Posted by u/def-pri-pub
Sesse__ · 3 days ago
And similarly, entire generations of programmers were never taught Horner's scheme. You can see it in the article, where they write stuff like

  A * x * x * x * x * x * x + B * x * x * x * x + C * x * x + D
(10 muls, 3 muladds)

instead of the faster

  tmp = x * x;
  ((A * tmp + B) * tmp + C) * tmp + D
(1 mul, 3 muladds)

arkmm · 3 days ago
Didn't know this technique had a name, but I would think a modern compiler could make this optimization on its own, no?
arkmm commented on Qwen3.5 Fine-Tuning Guide   unsloth.ai/docs/models/qw... · Posted by u/bilsbie
azath92 · 10 days ago
Only to prompt thought on this exact question, im interested in answers:

I just ran a benchmark against haiku of a very simple document classification task that at the moment we farm out to haiku in parallel. very naive same prompt system via same api AWS bedrock, and can see that the a few of the 4b models are pretty good match, and could be easily run locally or just for cheap via a hosted provider. The "how much data and how much improvement" is a question i dont have a good intuition for anymore. I dont even have an order of magnitude guess on those two axis.

Heres raw numbers to spark discussion:

| Model | DocType% | Year% | Subject% | In $/MTok |

|---------------|----------|-------|----------|-----------|

| llama-70b -----| 83 | 98 | 96 | $0.72 |

| gpt-oss-20b --| 83 | 97 | 92 | $0.07 |

| ministral-14b -| 84 | 100 | 90 | $0.20 |

| gemma-4b ----| 75 | 93 | 91 | $0.04 |

| glm-flash-30b -| 83 | 93 | 90 | $0.07 |

| llama-1b ------| 47 | 90 | 58 | $0.10 |

percents are doc type (categorical), year, and subject name match against haiku. just uses the first 4 pages.

in the old world where these were my own in house models, id be interested in seeing if i could uplift those nubmers with traingin, but i haven't done that with the new LLMs in a while. keen to get even a finger to the air if possible.

Can easily generate tens of thousands of examples.

Might try myself, but always keen for an opinion.

_edit for table formatting_

arkmm · 10 days ago
You can fine tune a small LLM with a few thousand examples in just a few hours for a few dollars. It can be a bit tricky to host, but if you share a rough idea of the volume and whether this needs to be real-time or batched, I could list some of the tradeoffs you'd think about.

Source: Consulted for a few companies to help them finetune a bunch of LLMs. Typical categorical / data extraction use cases would have ~10x fewer errors at 100x lower inference cost than using the OpenAI models at the time.

arkmm commented on Qwen3.5 Fine-Tuning Guide   unsloth.ai/docs/models/qw... · Posted by u/bilsbie
faxmeyourcode · 10 days ago
Especially for super constrained applications. I don't care if the language model that I use for my extremely specific business domain can solve PhD math or remember the works of Shakespeare. I'd trade all of that for pure task specific accuracy.
arkmm · 10 days ago
Can you share more details about your use case? The good applications of fine tuning are usually pretty niche, which tends to make people feel like others might not be interested in hearing the details.

As a result it's really hard to read about real-world use cases online. I think a lot of people would love to hear more details - at least I know I would!

arkmm commented on Payment fees matter more than you think   cuencahighlife.com/why-pa... · Posted by u/dxs
arkmm · 11 days ago
Payment fees are crazy when you think about them from the perspective of a merchant in a low margin business. E.g. in retail or restaurants, margins aren't much better than ~10%. If they didn't have to pay ~3% credit card fees, they'd have 30% more profit!
arkmm commented on Nano Banana 2: Google's latest AI image generation model   blog.google/innovation-an... · Posted by u/davidbarker
kevincox · 16 days ago
I actually think this was a good thing. Manipulating images incredibly convincingly was already possible but the cost was high (many hours of highly skilled work). So many people assumed that most images they were seeing were "authentic" without much consideration. By making these fake images ubiquitous we are forcing people to quickly learn that they can't believe what they see on the internet and tracking down sources and deciding who you trust is critically important. People have always said that you can't believe what you see on the internet, but unfortunately many people have managed without major issue ignoring this advice. This wave will force them to take that advice to heart by default.
arkmm · 16 days ago
I used to also have this optimistic take, but over time I think the reality is that most people will instead just distrust unknown online sources and fall into the mental shortcuts of confirmation bias and social proof. Net effect will be even more polarization and groupthink.
arkmm commented on The First Fully General Computer Action Model   si.inc/posts/fdm1/... · Posted by u/nee1r
nee1r · 19 days ago
Hey guys! I’m Neel, been holed up in our south park office for the past year working on model training. excited to share our research!

This is a preview of a very different type of computer use model—we train on the internet. Specifically we have 11 million hours of computer video stored on our storage cluster (previously shared https://news.ycombinator.com/item?id=45438496 !) and the model can work in 30 FPS. Since we match the fundamental form factor of computer-use, we can get our model to do CAD, browse websites, and even drive a car using arrow keys. I’m super excited to see what our model can do as we scale more, it's a fun frontier to work on (not language models :) ).

The team and I will be online responding to the comments, so drop any questions.

arkmm · 17 days ago
Get ready for the acquisition offers.
arkmm commented on Do not apologize for replying late to my email   ploum.net/2026-02-11-do_n... · Posted by u/validatori
arkmm · a month ago
Sorry Ploum, just getting a chance to read this now and comment. Great insights!
arkmm commented on Claude Code is your customer   calebjohn.xyz/blog/b2cc/... · Posted by u/mfbx9da4
CuriouslyC · a month ago
"something simpler/simpler approach" are terms I search for in evals because they almost always indicate the model going off the rails (assuming the input prompt was decent).
arkmm · a month ago
this is a really cool insight, going to use this on my team from now on!
arkmm commented on Raspberry Pi's New AI Hat Adds 8GB of RAM for Local LLMs   jeffgeerling.com/blog/202... · Posted by u/ingve
myrmidon · 2 months ago
Are there significant usecases for the really small LLMs right now (<10b distills and such)?

My impression so far was that the resulting models are unusably stupid, but maybe there are some specific tasks where they still perform acceptably?

arkmm · 2 months ago
They're still very good for finetuned classification, often 10-100x cheaper to run at similar or higher accuracy as a large model - but I think most people just prompt the large model unless they have high volume needs or need to self host.

u/arkmm

KarmaCake day139April 15, 2020View Original