Readit News logoReadit News
nialv7 commented on Why are anime catgirls blocking my access to the Linux kernel?   lock.cmpxchg8b.com/anubis... · Posted by u/taviso
debugnik · 5 days ago
That's not bypassing it, that's them finally engaging with the PoW challenge as intended, making crawling slower and more expensive, instead failing to crawl at all, which is more of a plus.

This however forces servers to increase the challenge difficulty, which increases the waiting time for the first-time access.

nialv7 · 5 days ago
Obviously the developer of Anubis thinks it is bypassing: https://github.com/TecharoHQ/anubis/issues/978
nialv7 commented on Why are anime catgirls blocking my access to the Linux kernel?   lock.cmpxchg8b.com/anubis... · Posted by u/taviso
nialv7 · 5 days ago
> Fuck AI scrapers, and fuck all this copyright infringement at scale.

Yes, fuck them. Problem is Anubis here is not doing the job. As the article already explains, currently Anubis is not adding a single cent to the AI scrappers' costs. For Anubis to become effective against scrappers, it will necessarily have to become quite annoying for legitimate users.

nialv7 commented on I Prefer RST to Markdown (2024)   buttondown.com/hillelwayn... · Posted by u/shlomo_z
nialv7 · 8 days ago
Uh, why haven't we drilled it into people's brains that regex cannot be used to parse matching parentheses/brackets?
nialv7 commented on Simulator of the life of a 30-year-old in the UK   nicksimulator.com/... · Posted by u/kostyal
nialv7 · 8 days ago
It's created by Progress, which IIUC is a movement under the Labour party?

I mean, you are literally in power, you can just change it. What's the point of this?

nialv7 commented on Woz: 'I Am the Happiest Person'   daringfireball.net/linked... · Posted by u/mariuz
nialv7 · 9 days ago
Too bad the world isn't run by people like him.
nialv7 commented on It seems like the AI crawlers learned how to solve the Anubis challenges   social.anoxinon.de/@Codeb... · Posted by u/moelf
nialv7 · 9 days ago
It is just sad we are in a time where measures like Anubis is necessary. The author's efforts are admirable, so I don't mean this personally: but Anubis is a bad product IMHO.

It doesn't quite do what it is advertised to do, as evidenced by this post; and it degrades user experience for everybody. And it also stops the website from being indexed by search engines (unless specifically configured otherwise). For example, gitlab.freedesktop.org pages have just disappeared from Google.

We need to find a better way.

nialv7 commented on Citybound: City building game, microscopic models to vividly simulate organism   aeplay.org/citybound... · Posted by u/modinfo
nialv7 · 11 days ago
needs a "(2020)" in the title. this is not an active project.
nialv7 commented on Training language models to be warm and empathetic makes them less reliable   arxiv.org/abs/2507.21919... · Posted by u/Cynddl
nialv7 · 13 days ago
Well, haven't we seen similar results before? IIRC finetuning for safety or "alignment" degrades the model too. I wonder if it is true that finetuning a model for anything will make it worse. Maybe simply because there is just orders of magnitudes less data available for finetuning, compared to pre-training.
nialv7 commented on GPT-5   openai.com/gpt-5/... · Posted by u/rd
Topfi · 18 days ago
> 400,000 context window

> 128,000 max output tokens

> Input $1.25

> Output $10.00

Source: https://platform.openai.com/docs/models/gpt-5

If this performs well in independent needle-in-haystack and adherence evaluations, this pricing with this context window alone would make GPT-5 extremely competitive with Gemini 2.5 Pro and Claude Opus 4.1, even if the output isn't a significant improvement over o3. If the output quality ends up on-par or better than the two major competitors, that'd be truly a massive leap forward for OpenAI, mini and nano maybe even more so.

nialv7 · 18 days ago
Being on-par with competitors is somehow a "massive leap" for OpenAI now? How far have they fallen...
nialv7 commented on GPT-5   openai.com/gpt-5/... · Posted by u/rd
surround · 18 days ago
GPT-5 knowledge cutoff: Sep 30, 2024 (10 months before release).

Compare that to

Gemini 2.5 Pro knowledge cutoff: Jan 2025 (3 months before release)

Claude Opus 4.1: knowledge cutoff: Mar 2025 (4 months before release)

https://platform.openai.com/docs/models/compare

https://deepmind.google/models/gemini/pro/

https://docs.anthropic.com/en/docs/about-claude/models/overv...

nialv7 · 18 days ago
maybe OpenAI have a terribly inefficient data ingestion pipeline? (wild guess) basically taking in new data is tedious so they do that infrequently and keep using old data for training.

u/nialv7

KarmaCake day1867April 23, 2014
About
Galactic Hitchhiker
View Original