Readit News logoReadit News
Vuizur commented on Gemma 3n preview: Mobile-first AI   developers.googleblog.com... · Posted by u/meetpateltech
ai-christianson · 4 months ago
That seems way too good to be true.

What's the catch?

Vuizur · 4 months ago
It is not very good at hard tasks, its ranking is much worse there.
Vuizur commented on The behavior of LLMs in hiring decisions: Systemic biases in candidate selection   davidrozado.substack.com/... · Posted by u/hunglee2
Vuizur · 4 months ago
The next question is if LLMs are actually more sexist than the average human working in HR. I am not so sure...
Vuizur commented on Show HN: Offline audiobook from any format with one CLI command   github.com/C-Loftus/Quick... · Posted by u/C-Loftus
leshokunin · a year ago
Would this dedrm my audible stuff?
Vuizur · a year ago
Vuizur commented on xAI's Colossus: Most Powerful AI Cluster Online in 122 Days   twitter.com/elonmusk/stat... · Posted by u/gfortaine
torlok · a year ago
Does xAI do anything interesting or are they just trying to catch up?
Vuizur · a year ago
Grok-2 is rank 2 on LLM arena, it's basically as good as the best Gemini model. They already caught up. Only the latest ChatGPT model is a tiny bit better.
Vuizur commented on Xapian: Open source search engine library   xapian.org/... · Posted by u/Bluestein
Vuizur · a year ago
I once wanted to compile a program that used Xapian on Windows. It was basically impossible for mortals.

Imo people should use cross-platform alternatives.

Vuizur commented on Ask HN: Predictions for when GPT-5 will be released and how safe it will be?    · Posted by u/Heidaradar
Vuizur · a year ago
It will likely be amazing, Sam Altman said that the step between 4 and 5 will be like the one between 3.5 and 4. You can of course doubt him, but we'll see...

I guess it will be this year, some guy working at OpenAI already posted "4+1=5" on Twitter, which is suggestive.

Vuizur commented on Wikimedia Enterprise – APIs for LLMs, AI Training, and More   enterprise.wikimedia.com/... · Posted by u/ks2048
Vuizur · a year ago
Wikimedia is unfortunately becoming one of the worst places to give your money to. They have their closed-source infrastructure, which now for years has not been able to generate HTML dumps without a significant percentage of articles missing.

They have know of the bug for ages, but still...

WMF don't care about their products at all, they give much more funding to vanity workshops in Africa while utterly ignoring the requests of the Wiktionary community, for example.

Vuizur commented on Replacing pyinstaller with 100 lines of code   tushar.lol/post/packaged/... · Posted by u/todsacerdoti
Vuizur · a year ago
There is also PyApp, which I think is really promising. The docs there are not that comprehensive yet and maybe a bit confusing, but the packaged programs usually work out of the box, unlike with pyinstaller.
Vuizur commented on GitHub Copilot Workspace: Technical Preview   github.blog/2024-04-29-gi... · Posted by u/davidbarker
krainboltgreene · a year ago
> Student here: I legitimately cannot understand how senior developers can dismiss these LLM tools

Because we've seen similar hype before and we know what impactful change looks like, even if we don't like the impact (See: Kubernetes, React, MongoDB).

> executing large-scale changes in entire repositories in 3 years

Is this actually happening? I haven't seen any evidence of that.

Vuizur · a year ago
>executing large-scale changes in entire repositories in 3 years

You can look at SWE-Agent, it solved 12 percent of the GitHub issues of their test dataset. It probably depends on your definition of large-scale.

This will get much better, it is a new problem with lots of unexplored details, and we will likely get GPT-5 this year, which is supposed to be a similar jump in performance as from 3.5 to 4 according to Altman.

Vuizur commented on ChatGPT provides false information about people, and OpenAI can't correct it   noyb.eu/en/chatgpt-provid... · Posted by u/skilled
bux93 · a year ago
Meh. GDPR sets limits on 'processing' personally identifiable information. In the context of an LLM, its outputs may contain PII if its inputs do. Those inputs are training input and prompts. So long(!) as the training input doesn't have PII, the output will only have it if the prompts do. Same as if you save a file on onedrive, if you save PII there, you're the data controller, and Microsoft is a processor on your behalf.
Vuizur · a year ago
It is impossible to remove personal data ("any information which are related to an identified or identifiable natural person") from the LLM training data.

As far as I understand it ChatGPT and all other similar systems are blatantly violating GDPR, they would have to for example publish their related training data to conform.

I guess the EU authorities don't do anything for now because they don't want to admit that their funny law basically bans all state-of-the-art AI.

(Ok, Openai also broke the law in almost all countries by downloading shadow libraries, but here they at least have more plausible deniability.)

u/Vuizur

KarmaCake day45January 1, 2023View Original