Readit News logoReadit News
lwneal commented on Auto-AVSR: Audio-Visual Speech Recognition with Automatic Labels   github.com/mpc001/auto_av... · Posted by u/yagizdegirmenci
lwneal · a year ago
Not referenced in the README, here's a great video demonstration of this type of AVSR network running in real time:

https://m.youtube.com/watch?v=XDO8OYnmkNY&t=120s

lwneal commented on Show HN: #!/usr/bin/env docker run   gist.github.com/adtac/595... · Posted by u/adtac
lwneal · 2 years ago
You can create this type of thing (a self-contained single-file project) for any language or infrastructure, with or without a clever shebang. All you need are heredocs.

For example, here's the same app but packaged as a regular bash script:

https://gist.github.com/lwneal/a24ba363d9cc9f7a02282c3621afa...

lwneal commented on Before Altman’s ouster, OpenAI’s board was divided and feuding   nytimes.com/2023/11/21/te... · Posted by u/vthommeret
lwneal · 2 years ago
The relevant passage from the paper co-written by board member Helen Toner:

"OpenAI has also drawn criticism for many other safety and ethics issues related to the launches of ChatGPT and GPT-4, including regarding copyright issues, labor conditions for data annotators, and the susceptibility of their products to "jailbreaks" that allow users to bypass safety controls...

A different approach to signaling in the private sector comes from Anthropic, one of OpenAI's primary competitors. Anthropic's desire to be perceived as a company that values safety shines through across its communications, beginning from its tagline: "an AI safety and research company." A careful look at the company's decision-making reveals that this commitment goes beyond words."

[1] https://cset.georgetown.edu/publication/decoding-intentions/

lwneal commented on GPT-4V(ision) system card [pdf]   cdn.openai.com/papers/GPT... · Posted by u/juunge
simonw · 2 years ago
Genuine question: why is this only published as a PDF?

OpenAI have the resources to also publish this as HTML. They chose not to.

They're not alone in this - most of the academic and research world, plus the concept of a "whitepaper" seems predicated on the idea of publishing PDFs.

Is this some stupid thing where human beings are expected to attach more prestige to information published in this way?

PDFs are a terrible way of publishing information in 2023:

- they render poorly on mobile devices, where many (most?) people do their reading

- they're hard to copy and paste information out of

- you can't link to headings within them (like HTML fragment links)

- you can't easily run them through translation tools like the one built into Chrome

The benefits of PDF I can see are:

1. Easier to print and get the exact expected output

2. You can save one file offline

3. Easier to author

I'm not arguing to replace PDFs with HTML (though I wouldn't miss them personally) - I'm saying publish documents as both!

Provide an HTML version and a PDF alternative for people who want it.

Am I missing something here? Why does the academic and research world stubbornly stick to such a hostile way of publishing their results?

lwneal · 2 years ago
I think it's about citation. Traditionally, a pdf is a complete and finished work, analogous to a published journal article or book. It is static content and will not change, unlike HTML which might be "under construction".

This isn't necessarily still true: HTML content can stay up on the web forever and a pdf can change, but people still prefer to cite something that looks like a paper document.

Since a whitepaper is often meant to be cited, it's published as a pdf to take advantage of this preference.

The best approach is to publish a PDF for citation along with a public HTML demo, like https://jonbarron.info/mipnerf360/

lwneal commented on Ts_zip: Text Compression Using Large Language Models   bellard.org/ts_server/ts_... · Posted by u/Deeg9rie9usi
lwneal · 3 years ago
A man goes to prison, and the first night while he's laying in bed, he hears someone yell out, "44!", followed by laughter from the other prisoners.

Puzzled, he laid back down, but then he heard someone else yell out, "72!", followed by even more laughter.

"What's going on?" he asked his cellmate.

"Well, we've all heard every joke so many times, we've given them each a number to make it easier."

"Oh," he says, "can I try?"

"Sure, go ahead."

So, he yells out "102!" and the place goes nuts. People are whooping and laughing in a hysteria. He looks at his cellmate rolling on the ground laughing.

"Wow, good joke, huh?"

"Yeah! We ain't never heard that one before!"

lwneal commented on Nvidia DGX GH200 Whitepaper   resources.nvidia.com/en-u... · Posted by u/volta87
wmf · 3 years ago
A wafer of H100s uses far more electricity than a wafer of [Apple] A16s though.
lwneal · 3 years ago
An H100 uses up to 350 Watts, while an A16 has a TDP of only 8 W. But, the A16 is a smaller chip (about 108mm vs. the H100's 814mm) so you can fit more of them on a wafer. Since a wafer is 300mm in diameter, its area is 70685 mm^2, which would yield 86 H100's or 654 A16's. [1][2]

However, that discounts the waste on the edges of the circular wafer, as well as the chip yield, which will both likely be worse for the larger chip [3]. But, assuming a generous 70% yield by area [4], one wafer's worth of H100s all packaged into GPUs and running full blast will use maybe 20 kilowatts, while the same wafer of A16s might use 3.6 kilowatts. Although in practice, the A16s will spend most of their time conserving battery power in your pocket, and even the H100s will spend some of their time idle.

TSMC is now producing over 14 million wafers per year. At most 1.2 million of those are on the 3nm node, and not all of that production goes to GPUs. But as an upper bound, if we imagine that all of TSMC's wafers could be filled up with nothing but H100 chips, and if all of those H100 chips were immediately put to use running AI 24/7, how much additional load could it put on the power grid every year?

The answer is, around 280 gigawatts, or if they were running 24/7 for a year, about 2500 terawatt-hours. That's about 10% of current world electricity consumption! So it's not completely implausible to imagine that a huge ramp-up in AI usage might have an effect on the electric grid.

*edit: This assumes we're talking about the Apple A16 (ie. the difference between phone chips and GPU chips). If we're talking about the Nvidia A16 (ie. the difference between current GPU chips and last node's GPU chips) see pclmulqdq's comment. ⠀

[1] https://nanoreview.net/en/soc/apple-a16-bionic

[2] https://www.techpowerup.com/gpu-specs/h100-pcie-80-gb.c3899

[3] https://news.ycombinator.com/item?id=24185108

[4] https://www.extremetech.com/computing/analyst-tsmc-hitting-5...

[5] https://www.tsmc.com/english/dedicatedFoundry/manufacturing/...

[6] https://www.wolframalpha.com/input?i=%2814+million%29+*+%282...*

lwneal commented on AI Canon   a16z.com/2023/05/25/ai-ca... · Posted by u/nihit-desai
lwneal · 3 years ago
This is a fine list, but it only covers a specific type of generative AI. Any set of resources about AI in general has to at least include the truly canonical Norvig & Russel textbook [1].

Probably also canonical are Goodfellow's Deep Learning [2], Koller & Friedman's PGMs [3], the Krizhevsky ImageNet paper [4], the original GAN [5], and arguably also the AlphaGo paper [6] and the Atari DQN paper [7].

[1] https://aima.cs.berkeley.edu/

[2] https://www.deeplearningbook.org/

[3] https://www.amazon.com/Probabilistic-Graphical-Models-Princi...

[4] https://proceedings.neurips.cc/paper_files/paper/2012/file/c...

[5] https://arxiv.org/abs/1406.2661

[6] https://www.nature.com/articles/nature16961

[7] https://www.nature.com/articles/nature14236

u/lwneal

KarmaCake day1589January 12, 2020
About
https://lwneal.com
View Original