Readit News logoReadit News
slaucon commented on Failing to Understand the Exponential, Again   julian.ac/blog/2025/09/27... · Posted by u/lairv
slaucon · 3 months ago
I feel like there should be some take away from the fact that we have to come up with new and interesting metrics like “Length of a Task That Can Be Automated” in order to point out that exponential growth is still happening. Fwiw, it does seem like a good metric, but it also feels like you can often find some metric that’s improving exponentially even when the base function is leveling out.
slaucon commented on Show HN: A resume filtering puzzle running on a VM running in browser via WASM   treeline.ai/puzzle/... · Posted by u/hussainkader
slaucon · 7 months ago
Absolutely insane that the emulator, the kernel, AND the os are just 24mb combined.
slaucon commented on Parasitic worms 'manipulate' mantises onto asphalt roads, say researchers   mainichi.jp/english/artic... · Posted by u/anigbrowl
slaucon · a year ago
In a similar vein, I always found it interesting (although frightening) that rabies cause hydrophobia. The theory is that drinking water can wash away the virus from your saliva, inhibiting its ability to spread through bites.

It makes sense that a virus passed through saliva would evolve like this, but I just find it particularly unsettling when a pathogen can effect higher-level behaviors like drinking water (or jumping into water for mantises).

slaucon commented on Why are cancer guidelines stuck in PDFs?   seangeiger.substack.com/p... · Posted by u/huerne
upghost · a year ago
It's so much worse than you could possibly imagine. I worked for a healthcare startup working on patient enrollment for clinical oncology trials. The challenges are amazing. Quite frankly it wouldn't matter if the data were in plaintext. The diagnostic codes vary between providers, the semantic understanding of the diagnostic information has different meanings between providers, electronic health records are a mess, things are written entirely in natural language rather than some kind of data structure. Anyone who's worked in healthcare software can tell you way more horror stories.

I do hope that LLMs can help straighten some of it out but anyone whos done healthcare software, the problems are not technical, they are quite human.

That being said one bright spot is we've (my colleagues, not me) made a huge step forward using category theory and Prolog to discover the provably optimal 3+3 clinical oncology dose escalation trial protocol[1]. David gave a great presentation on it at the Scryer Prolog meetup[2] in Vienna.

It's kind of amazing how in the dark ages we are with medicine. Even though this is the first EXECUTABLE/PROGRAMMABLE SPEC for a 3+3 cancer trial, he is still fighting to convince his medical colleagues and hospital administrators that this is the optimal trial because -- surprise -- they don't speak software (or statistics).

[1]: https://arxiv.org/abs/2402.08334

[2]: https://www.digitalaustria.gv.at/eng/insights/Digital-Austri...

slaucon · a year ago
This is a fascinating idea!
slaucon commented on Why are cancer guidelines stuck in PDFs?   seangeiger.substack.com/p... · Posted by u/huerne
pcrh · a year ago
The fundamental idea here is that doctors find it difficult to ensure that their recommendations are actually up-to-date with the latest clinical research.

Further, that by virtue of being at the centre of action in research, doctors in prestige medical centres have an advantage that could be available to all doctors. It's a pretty important point, sometimes referred to as the dissemination of knowledge problem.

Currently, this is best approached by publishing systematic reviews according to the Cochrane Criteria [0]. Such reviews are quite labour-intensive and done all too rarely, but are very valuable when done.

One aspect of such reviews, when done, is how often they discard published studies for reasons such as bias, incomplete datasets, and so forth.

The approach described by Geiger in the link is commendable for its intentions but the outcome will be faced with the same problem that manual systematic reviews face.

I wonder if the author considered included rules-based approaches (e.g. Cochrane guidelines) in addition to machine learning approaches?

[0] https://training.cochrane.org/handbook

slaucon · a year ago
Hey author here--Cochrane reviews are great.

NCCN guidelines and Cochrane Reviews serve complementary roles in medicine - NCCN provides practical, frequently updated cancer treatment algorithms based on both research and expert consensus, while Cochrane Reviews offer rigorous systematic analyses of research evidence across all medical fields with a stronger focus on randomized controlled trials. The NCCN guidelines tend to be more immediately applicable in clinical practice, while Cochrane Reviews provide a deeper analysis of the underlying evidence quality.

My main goal here was to show what you could do with any set of medical guidelines that was properly structured. You can choose any criteria you want.

slaucon commented on Why are cancer guidelines stuck in PDFs?   seangeiger.substack.com/p... · Posted by u/huerne
prepend · a year ago
I’d rather have the pdf than a custom tool. Especially considering the tool will be unique to the practice or emr. And likely expensive to maintain.

PDFs suck in many ways but are durable and portable. If I work with two oncologists, I use the same pdf.

The author means well but his solution will likely be worse because only he will understand it. And there’s a million edge cases.

slaucon · a year ago
Hey author here! Appreciate the feedback! Agreed on importance of portability and durability.

I'm not trying to build this out or sell it as a tool to providers. Just wanted to demo what you could do with structured guidelines. I don't think there's any reason this would have to be unique to a practice or emr.

As sister comments mentioned, I think the ideal case here would be if the guideline institutions released the structured representations of the guidelines along with the PDF versions. They could use a tool to draft them that could export in both formats. Oncologists could use the PDFs still, and systems could lean into the structured data.

slaucon commented on Ask HN: Stanford CS 153 help    · Posted by u/anjneymidha
slaucon · a year ago
A progression of projects that comes to mind:

1) CI and IAC that deploy a web app running in a container

2) Add horizontal scaling and load balancer

3) Add long running tasks / scheduled task support

4) Deploys will likely break long running tasks. Implement blue/green or rolling deploys or some other sort of advanced deployment scheme

5) Implement rollbacks

slaucon commented on Show HN: Outerbase Studio – Open-Source Database GUI   github.com/outerbase/stud... · Posted by u/burcs
carlosjobim · a year ago
I remember I used TablePlus, which was what you described. Very pleasant program. Not browser based, though.
slaucon · a year ago
Yeah, my last company paid for a subscription to this. Enjoyed using it. Don’t think there’s a massive market, but definitely lots of devs who want easy DB access and would pay $5/month.
slaucon commented on What happens if we remove 50 percent of Llama?   neuralmagic.com/blog/24-s... · Posted by u/BUFU
slaucon · a year ago
> “By sourcing and filtering only the highest-quality and most representative data for LLM use cases, we reduced the pretraining set to just 13 billion tokens—drastically cutting the environmental impact of further training while preserving performance.”

Would love to know more about how they filtered the training set down here and what heuristics were involved.

I think that the models we use now are enormous for the use cases we’re using them for. Work like this and model distillation in general is fantastic and sorely needed, both to broaden price accessibility and to decrease resource usage.

I’m sure frontier models will only get bigger, but I’d be shocked if we keep using the largest models in production for almost any use case.

u/slaucon

KarmaCake day79October 21, 2022View Original