Readit News logoReadit News
rythie commented on No AI* Here – A Response to Mozilla's Next Chapter   waterfox.com/blog/no-ai-h... · Posted by u/MrAlex94
rythie · a day ago
Waterfox is dependant on Firefox still being developed. Mozilla are adding these features to try to stay relevant and keep or gain market share. If this fails, and Firefox goes away, Waterfox is unlikely to survive.
rythie commented on Ask HN: How can ChatGPT serve 700M users when I can't run one GPT-4 locally?    · Posted by u/superasn
rythie · 4 months ago
First off I’d say you can run models locally at good speed, llama3.1:8b runs fine a MacBook Air M2 with 16GB RAM and much better on a Nvidia RTX3050 which are fairly affordable.

For OpenAI, I’d assume that a GPU is dedicated to your task from the point you press enter to the point it finishes writing. I would think most of the 700 million barely use ChatGPT and a small proportion use it a lot and likely would need to pay due to the limits. Most of the time you have the website/app open I’d think you are either reading what it has written, writing something or it’s just open in the background, so ChatGPT isn’t doing anything in that time. If we assume 20 queries a week taking 25 seconds each. That’s 8.33 minutes a week. That would mean a single GPU could serve up to 1209 users, meaning for 700 million users you’d need at least 578,703 GPUs. Sam Altman has said OpenAI is due to have over a million GPUs by the end of year.

I’ve found that the inference speed on newer GPUs is barely faster than older ones (perhaps it’s memory speed limited?). They could be using older clusters of V100, A100 or even H100 GPUs for inference if they can get the model to fit or multiple GPUs if it doesn’t fit. A100s were available in 40GB and 80GB versions.

I would think they use a queuing system to allocate your message to a GPU. Slurm is widely used in HPC compute clusters, so might use that, though likely they have rolled their own system for inference.

rythie commented on The QR Backlash Has Won. Restaurants Are Ditching Them for Good.   wsj.com/business/hospital... · Posted by u/bookofjoe
tootie · 2 years ago
I much prefer the QR codes. I can zoom and search and turn up the brightness for my old eyes. It's funny seeing the current generation get up in arms about progress after spending so much time whining about boomers inability to cope with change.
rythie · 2 years ago
We know how to use the apps, we just don’t want to. Not all change is for the better.
rythie commented on IMAX emulates PalmPilot software to power Oppenheimer’s 70 mm release   arstechnica.com/gadgets/2... · Posted by u/riffraff
sillywalk · 2 years ago
I wonder why they chose Palm, but then why not, and what else would they use? Ipaq?

It's not clear, but I assume it sends commands to the actual film hardware, and its not doing some real-time control.

"The software shows a handful of controls for the projectionist to queue up the film and control the platters that feed film at six feet per second. " [0]

[0] https://www.extremetech.com/mobile/imax-using-20-year-old-pa...

rythie · 2 years ago
Palm was the market leader, it would have been the obvious choice. Palm had been around since 1996 and by 1998 had sold 30 million devices [1]. PocketPC didn’t come out until 2000, in 2001 they had only sold 1.25 million devices, equating to less than 10% market share [2]. From what I remember Palm Pilots were the go to choice for PDAs, they were simple and worked. Other devices had come and gone. It would have been odd if they chosen something else. I doubt anyone was thinking it would be used for 20 years, though I don’t think people would have thought it would go away at the time.

[1] https://history-computer.com/palm-pilot-guide/ [2] https://www.zdnet.com/article/pocket-pc-sales-1-million-and-...

rythie commented on Ask HN: Worth it to buy 4x Nvidia Tesla K40 for AI?    · Posted by u/speedylight
rythie · 3 years ago
The pytorch binaries from pip and conda won’t work on these GPUs, though there are some alternative binaries being maintained that still work: https://blog.nelsonliu.me/2020/10/13/newer-pytorch-binaries-...

The latest Nvidia driver no longer supports the K40, so you’ll have to use version 470 (or lower, officially Nvidia says 460, but 470 seems to work). That supports CUDA 11.4 natively. Newer versions of CUDA 11.x are supported: https://docs.nvidia.com/deploy/cuda-compatibility/index.html though CUDA 12 is not.

In my testing, a system with a single RTX3060 was faster in tensorflow than with 3 K40s and probably close to the performance of 4 k40s.

If you are considering other GPUs, there are some good benchmarks here (The RTX3060 is not there, though the GTX1080Ti was almost the same performance in the tensorflow test they run): https://lambdalabs.com/gpu-benchmarks

As others have said Google CoLab is free option you can use.

rythie commented on AMD Killed the Itanium (2005)   utcc.utoronto.ca/~cks/spa... · Posted by u/todsacerdoti
_ph_ · 3 years ago
Of course the whole story is way more complex, but I have been saying for years that the Itanium might have succeeded, if AMD had not extended x86 to 64 bit. The 64 bit extensions did not only fix some of the x86 problems (it increased the register count, pushed 64 bit double floats) but it made x86 a choice for the more serious compute platforms and servers.

Back then, the whole professional world had switched to 64 bit, both from a performance and memory size perspective. That is why the dotcom time basically was based on Sparc Suns. The Itanium was way late, it still was Intels only offering in the domain. Until x86-64 came and very quickly entered the professional compute centers. The performance race in the consumer space then sealed the deal by providing faster CPUs than the classic RISC processors of the time, including Itanium.

It is a bit sad to see it go, I wonder how well the architecture would have performed in modern processes. After all, an iPhone has a much larger transistor count than those "large" and "hot" Itaniums.

rythie · 3 years ago
The industry was already moving away from the big 64 bit SMP machines made Sun, SGI & IBM. In many cases a cluster of 32bit x86 machines made more sense than one expensive big machine with high priced support contracts and parts. 32 bit x86 machines already supported more than 4GB total memory with PAE, it was just that one process couldn’t use more than 4GB. Other 64bit chips were already well established (SPARC, POWER, MIPS), probably for most of the users they couldn’t easily move to a new CPU architecture. For other users by the time they needed the bigger machines x86 64bit was already available, including from Intel themselves. AMD was limited 8 sockets from what I remember, so their was still a small market for big Itanium systems (like SGI’s Altix).
rythie commented on Smartphones wiped out 97% of the compact camera market   asia.nikkei.com/Business/... · Posted by u/joos3
adav · 3 years ago
Which model is considered by photographers as the last "best" compact camera?
rythie · 3 years ago
The Sony RX100 series would be in most top lists. Though personally I'm not sure why they went with a slower lens from the mark 6 onwards. The ZV1 continues with the a similar lens from earlier models. I have the RX100 mark iii, that's still quite good.
rythie commented on Minikube now supports rootless podman driver for running Kubernetes   github.com/kubernetes/min... · Posted by u/encryptluks2
coffeekid · 3 years ago
You can also call docker commands by being part of the docker group IIRC.

Doesn't this have more to do with the daemon that the user executing commands ?

rythie · 3 years ago
With podman there is no daemon, everything is running as you. The standard setup for docker has a daemon running as root, which means when you start a container it has root privileges.
rythie commented on Minikube now supports rootless podman driver for running Kubernetes   github.com/kubernetes/min... · Posted by u/encryptluks2
srvmshr · 3 years ago
Please excuse my ignorance. I am aware of how Docker generally operates (at a noob level i.e. containerizing an application & uploading to container public registry for general availability).

Given that understanding, could someone please explain what "rootless" would mean? I want to understand these in simpler terms:)

(Thank you in advance)

rythie · 3 years ago
It means you don't need to be root to run it.
rythie commented on Three Electric-Jaguar Years   tbray.org/ongoing/When/20... · Posted by u/CapitalistCartr
michaelt · 4 years ago
> Slow charge infrastructure is much cheaper to build out.

At a highway rest stop, why would a fast charger that can charge 1 car in 6 minutes be any more expensive than a bank of slow chargers that can charge 10 cars in 60 minutes?

rythie · 4 years ago
A fast charger has AC-DC converter provide DC power to car, which is expensive. Slow chargers are AC, which are very cheap in comparison.

u/rythie

KarmaCake day3500June 22, 2008
About
Contact Info:

   Blog: https://blog.rythie.com
   Twitter: https://twitter.com/rythie

View Original