Readit News logoReadit News
jroesch commented on Show HN: Cuq – Formal Verification of Rust GPU Kernels   github.com/neelsomani/cuq... · Posted by u/nsomani
nsomani · 2 months ago
That instinct is right. cuTile would be easier to parse but harder to reason about formally.
jroesch · 2 months ago
We also have a formal memory model and the program semantics are simpler so if anything reasoning about it should be easier.
jroesch commented on Nvidia announces next-gen RTX 5090 and RTX 5080 GPUs   theverge.com/2025/1/6/243... · Posted by u/somebee
PaulKeeble · a year ago
Looks like most of the improvement is only going to come when DLSS4 is in use and its generating most of the frame for Ray Tracing and then also generating 3 predicted frames. When you use all that AI hardware then its maybe 2x, but I do wonder how much fundamental rasterisation + shaders performance gain there is in this generation in practice on the majority of actual games.
jroesch · a year ago
There was some solid commentary on the Ps5Pro tech talk stating core rendering is so well optimized much of the gains in the future will come from hardware process technology improvements not from radical architecture changes. It seems clear the future of rendering is likely to be a world where the gains come from things like dlss and less and free lunch savings due to easy optimizations.
jroesch commented on Making AMD GPUs competitive for LLM inference (2023)   blog.mlc.ai/2023/08/09/Ma... · Posted by u/plasticchris
jroesch · a year ago
Note: this is old work, and much of the team working on TVM, and MLC were from OctoAI and we have all recently joined NVIDIA.
jroesch commented on Launch HN: Deepsilicon (YC S24) – Software and hardware for ternary transformers    · Posted by u/areddyyt
danjl · a year ago
In my experience, trying to switch VFX companies from CPU-based rendering to GPU-based rendering 10+ years ago, a 2-5x performance improvement wasn't enough. We even provided a compatible renderer that accepted Renderman files and generated matching images. Given the rate of improvement of standard hardware (CPUs in our case, and GPU-based inference in yours), a 2-5x improvement will only last a few years, and the effort to get there is large (even larger in your case). Plus, I doubt you'll be able to get your HW everywhere (i.e. mobile) where inference is important, which means they'll need to support their existing and your new SW stack. The other issue is entirely non-technical, and may be an even bigger blocker -- switching the infrastructure of a major LLM provider to a new upstart is just plain risky. If you do a fantastic job, though, you should get aquahired, probably with a small individual bonus, not enough to pay off your investors.
jroesch · a year ago
Having been working in DL inference for now 7+ years (5 of which at startup) which makes me comparably ancient in the AI world at this point. The performance rat race/treadmill is never ending, and to your point a large (i.e 2x+) performance improvement is not enough of a "painkiller" for customers unless there is something that is impossible for them to achieve without your technology.

The second problem is distribution: it is already hard enough to obtain good enough distribution with software, let alone software + hardware combinations. Even large silicon companies have struggled to get their HW into products across the world. Part of this is due to the actual purchase dynamics and cycle of people who buy chips, many design products and commit to N year production cycles of products built on certain hardware SKUs, meaning you have to both land large deals, and have opportune timing to catch them when they are evening shopping for a new platform. Furthermore the people with existing distribution i.e the Apple, Google, Nvidia, Intel, AMD, Qualcomms of the world already have distribution and their own offerings in this space and will not partner/buy from you.

My framing (which has remained unchanged since 2018) is that for silicon platform to win you have to beat the incumbents (i.e Nvidia) on the 3Ps: Price (really TCO), Performance, and Programmability.

Most hardware accelerators may win on one, but even then it is often theoretical performance because it assumes their existing software can/will work on your chip, which it often doesn't (see AMD and friends).

There are many other threats that come in this form, for example if you have a fixed function accelerator and some part of the model code has to run on CPU the memory traffic/synchronization can completely negate any performance improvements you might offer.

Even many of the existing silicon startups have been struggling with this since them middle of the last decade, the only thing that saved them is the consolidation to Transformers but it is very easy for a new model architecture to come out and require everyone to rework what they have built. This need for flexibility is what has given rise to the design ethos around GPGPU as flexibility in a changing world is a requirement not just a nice to have.

Best of luck, but these things are worth thinking deeply about as when we started in this market we were already aware of many of these things but their importance and gravity in the AI market have only become more important, not less :)

jroesch commented on AI models collapse when trained on recursively generated data   nature.com/articles/s4158... · Posted by u/rntn
simonw · a year ago
I get the impression that scraping the web isn't nearly as important a source of LLM training data as it used to be.

Everyone is trimming down their training data based on quality - there are plenty of hints about that in the Llama 3.1 paper and Mistral Large 2 announcement.

OpenAI are licensing data from sources like the Associated Press.

Andrej Karpathy said this: https://twitter.com/karpathy/status/1797313173449764933

> Turns out that LLMs learn a lot better and faster from educational content as well. This is partly because the average Common Crawl article (internet pages) is not of very high value and distracts the training, packing in too much irrelevant information. The average webpage on the internet is so random and terrible it's not even clear how prior LLMs learn anything at all.

jroesch · a year ago
I think this is roughly correct. My 2c is that folks used the initial web data to cold start and bootstrap the first few models, but so much of the performance increase we have seen at smaller sizes is a shift towards more conscientious data creation/purchase/curation/preparation and more refined evaluation datasets. I think the idea of scraping random text except maybe for the initial language understanding pre-training phase will be diminished over time.

This is understood in the academic literature as well, as people months/years ago were writing papers that a smaller amount of high quality data, is worth more than a large amount of low quality data (which tracks with what you can pick up from an ML 101 education/training).

jroesch commented on Show HN: Kimchi Reader – Immersive Korean Learning with a Popup Dictionary   kimchi-reader.app... · Posted by u/alaanor
voussoir · 2 years ago
I absolutely disagree. Romanized Korean is harder to read than Korean itself. The sounds implied by the roman spelling do not match the sounds you're supposed to make, and you lose the syllable boundaries that aid in pronunciation. I feel like a caveman stumbling over my letters when I try to read a romanized word.

A one-page alphabet reference chart would be enough to remind the reader which letter is which without relying on the romanization crutch.

Normally I don't like to make argumentative internet comments but I really passionately think romanization is a detriment to a learning tool.

jroesch · 2 years ago
I agree with this after a short while I turned off the romanization in many learning apps as it just messes with/undermines your actual learning.
jroesch commented on Apple Tests ‘Apple GPT,’ Develops Generative AI Tools to Catch OpenAI   bloomberg.com/news/articl... · Posted by u/helsinkiandrew
hn_throwaway_99 · 2 years ago
> But the question is whether they are committed enough to AI to do it, and whether they have the required ML talent.

Given how business leaders throughout tech feel that AI is going to be transformative, I don't think commitment is really going to be a problem. Many leaders feel that "you either get good at AI or you don't exist in 10 years".

In terms of attracting talent, there are 3 main things top AI folks look for:

1. Money (they are people after all)

2. The infrastructure (both hardware and people/organization-wise) to support large AI projects.

3. The willingness to release these AI projects to a large swath of people (to have "impact" as folks like to say).

E.g. Google had 1 and 2 but their reticence to release their models and corporate infighting made many of the top Google researchers leave for gigs elsewhere. I think it remains to be seen how Apple will handle #3 as well.

jroesch · 2 years ago
To chime in Apple already has a lot of great ML talent they are just far more deliberate and slow to change their products. People forget that FaceID was/is one of the most cutting edge ML features ever developed/deployed when it was released a few years ago.

Siri is sort of a red herring because its built by teams and tech that existed before Apple acquired most of its ML talent and some of its inability to evolve has been due to internal politics not the inability to build tech. iOS 17 is an example of Apple moving towards more deep learning speech/text work. I would bet heavily we will see them catch up with well integrated pieces as they have Money, infra, and already the ability to go wide (i.e all iOS users, again think FaceID).

jroesch commented on Apple Tests ‘Apple GPT,’ Develops Generative AI Tools to Catch OpenAI   bloomberg.com/news/articl... · Posted by u/helsinkiandrew
smoldesu · 2 years ago
I don't think it's a very huge or exploitable niche for two reasons:

- Android, Windows, Linux and MacOS can already run local and private models just fine. Getting something product-ready for iPhone is a game of catch-up, and probably a losing battle if Apple insists on making you use their AI assistant over competing options.

- The software side needs more development. The current SOTA inferencing techniques for Apple Silicon in llama.cpp are cobbled together with Metal shaders and NEON instructions, neither of which are ideal or specific to Apple hardware. If Apple wants to differentiate their silicon from the 2016 Macbooks running LLaMA with AVX, then they have to develop CoreML's API further.

jroesch · 2 years ago
Things are already possible on today's hardware, see https://github.com/mlc-ai/mlc-llm which allows many models to be run on M1/M2 Macs, WASM, iOS and more. The main limiting factor will be small enough, high quality enough models that performance is high enough ultimately this is HW limited and they will need to improve the neural engine/map more computation on to it to make the mobile exp. possible.
jroesch commented on Looming demise of the 10x developer – an era of enthusiast programmers is ending   blog.testdouble.com/posts... · Posted by u/rrampage
justinko · 2 years ago
No, “treatment” is going to consist of stimulant medication, which long-term, will decrease quality of life.
jroesch · 2 years ago
This is also just straight up FUD. ADHD is one of the few psychiatric conditions that has numerous effective medications which work reliably for a large part of the effected population.

Stimulants work for a large number of people diagnosed with ADHD with very little negative effects and are safe modulo a few exceptions for long term use.

Some individuals have negative experiences with Stimulant medications but I know from personal experience and from many friends in the ADHD community stimulants have literally been life saving for them.

They don't just reach for them because they are out to get you but they are effective for many people.

Furthermore many people who choose to forgo medication develop lifestyle and substance use issues which negative effects far out weigh low dose stimulants.

As other commenters said they are just a tool you still have to work on interventions, behavior modification, and so on.

At the end of the day in many ways ADHD is a disability (even if sometimes a super power) and you can't just delete it with a prescription.

Even if you forgo meds there are so many ways to boost your attention and quality of life and lots of research on what is effective, treatment can be much more than just medication.

jroesch commented on Looming demise of the 10x developer – an era of enthusiast programmers is ending   blog.testdouble.com/posts... · Posted by u/rrampage
keyle · 2 years ago
Well since you call my name, I'm just saying: any time someone complains about focus or learning on the internet, here comes the old "probably ADHD" guy/gal.

The internet is making everyone productivity addicts, can't just live anymore and be different.

If the author does not bring ADHD in the conversation, I think it's rude to find him a condition on the back of his post, which is far more interesting than isolating a paragraph and pulling a diagnosis on him.

It's a great blog post.

jroesch · 2 years ago
It has nothing to do with the blog post quality or being different the guy has multiple key sentences which map to key ADHD experiences/symptoms. For those of us living with ADHD its just an empathy response as many of us have suffered from experiences which closely map to what he described in that paragraph. We are often just looking to share as many of us have improved our lives substantially after someone suggesting we should get ourselves checked out.

On this part in particular while it can be great to deeply follow your passion with extreme focus; pursuing things regardless of their importance in your overall life and at the cost of other interests, relationships, or responsibilities can be an empty and unfulfilling existence in the end. Furthermore life can be markedly better with the correct interventions and treatment.

People seemingly get offended by even a suggestion because many people have extreme stigma against conditions like ADHD as well as a lot of misinformation from people have very little understanding of the actual traits, diagnostic criteria, treatment and prevalence of it.

u/jroesch

KarmaCake day390December 22, 2011
About
[ my public key: https://keybase.io/jroesch; my proof: https://keybase.io/jroesch/sigs/2j48sumXwtl6--d-yVJAmjTVkhifaVZONdzHa413srU ]
View Original