Readit News logoReadit News
tedivm · 3 years ago
I love the full stack deep learning crew, and took their course several years ago in Berkeley. I highly recommend it.

One thing that always blows my mind is how much it is just not worth it to train LLMs in the cloud if you're a startup (and probably even less so for really large companies). Compared to 36 month reserved pricing the break even point was 8 months if you bought the hardware and rented out some racks at a colo, and that includes the on hands support. Having the dedicated hardware also meant that researchers were willing to experiment more when we weren't doing a planned training job, as it wouldn't pull from our budget. We spent a sizable chuck of our raise on that cluster but it was worth every penny.

I will say that I would not put customer facing inference on prem at this point- the resiliency of the cloud normally offsets the pricing, and most inference can be done with cheaper hardware than training. For training though you can get away with a weaker SLA, and the cloud is always there if you really need to burst beyond what you've purchased.

ericpauley · 3 years ago
> Compared to 36 month reserved pricing the break even point was 8 months if you bought the hardware and rented out some racks at a colo, and that includes the on hands support. Having the dedicated hardware also meant that researchers were willing to experiment more when we weren't doing a planned training job, as it wouldn't pull from our budget.

That’s having it both ways, of course. You can’t both recoup the hardware cost in 8 months and have “free” downtime.

Under this pricing you need at least 25%ish duty cycle to break even (in 3 years) so probably still favoring buying, but for some people that might not add up. Pricing also varies drastically between providers, so this may depend on choice there.

rfoo · 3 years ago
The thing is, this is compared to 36 months reserved instance, which also assumes 100% usage. For true on-demand you pay much more.
moonchrome · 3 years ago
Reserved pricing is not on-demand.
bradleyjg · 3 years ago
It’s the first widespread thing where build your own makes sense in a while. Most prior workflows either needed high SLAs, were bursty, or just didn’t add up to much.
fpgaminer · 3 years ago
My experience with many of these services renting mostly A100s:

LambdaLabs: For on-demand instances, they are the cheapest available option. Their offering is straightforward, and I've never had a problem. The downside is that their instance availability is spotty. It seems like things have gotten a little better in the last month, and 8x machines are available more often than not, but single A100s were rarely available for most of this year. Another downside is lack of persistent storage, meaning you have to transfer your data every time you start a new instance. They have some persistent storage in beta, but it's effectively useless since it's only in one region and there's no instances in that region that I've seen.

Jarvis: Didn't work for me when I tried them a couple months ago. The instances would never finish booting. It's also a pre-paid system, so you have to fill up your "balance" before renting machines. But their customer service was friendly and gave me a full refund so shrug.

GCP: This is my go-to so far. A100s are $1.1/hr interruptible, and of course you get all the other Google offerings like persistent disks, S3, managed SQL, container registry, etc. Availability of interruptible instances has been consistently quite good, if a bit confusing. I've had some machines up for a week solid without interruption, while other times I can tear down a stack of machines and immediately request a new one only to be told they are out of availability. The downsides are the usual GCP downsides: poor documentation, sometimes weird glitches, and perhaps the worst billing system I've seen outside of the healthcare industry.

Vast.ai: They can be a good chunk cheaper, but at the cost of privacy, security, support, and reliability. Pre-load only. For certain workloads and if you're highly cost sensitive this is a good option to consider.

RunPod: Terrible performance issues. Pre-load only. Non-responsive customer support. I ended up having to get my credit card company involved.

Self-hosted: As a sibling comment points out, self hosting is a great option to consider. In particular "Having the dedicated hardware also meant that researchers were willing to experiment more". I've got a couple cards in my lab that I use for experimentation, and then throw to the cloud for big runs.

pavelstoev · 3 years ago
please consider CoreWeave - great experience so far.
fpgaminer · 3 years ago
For 2 to 3 times the cost of Lambda on demand and GCP preempt?
sorcer · 3 years ago
+1 to this
pavelstoev · 3 years ago
May also suggest the suite of open source DeepView tools and which are part of PyPi. Profile and predict your specific model training performance on a variety of GPUs. I wrote a linked in post with usage GIF here: https://www.linkedin.com/posts/activity-7057419660312371200-...

And links to PyPi:

=> https://pypi.org/project/deepview-profile/ => https://pypi.org/project/deepview-predict/

And you can actually do it in browser for several foundational models (more to come): => https://centml.ai/calculator/

Note: I have personal interests in this startup.

stuckkeys · 3 years ago
"Those god damn AWS charges" -Silicon Valley. Might as well build your own GPU farm. Some of these cards, used you can probably get for 6K (guestimating).
x-complexity · 3 years ago
That would imply that the current AI cycle would be able to persist at its current levels of frothiness indefinitely: In the in-between lull periods, these GPU farms would be seen as something to sell off. This doesn't even take into account the eventual depreciation of the GPUs in question, as better GPUs/accelerators come into the market.

Most companies have an AWS account that they can throw on more money at for 'AI research & implementation'. With such an account existing in the first place, along with said price depreciations, the company in question would have to be certain that they'll use said GPUs all the time to make up for the upfront costs they'll be putting up with.

mirekrusin · 3 years ago
Your own hardware can be rented out if you're not using it through vast.ai for example.

When new hardware comes out, you can sell old one to recover some of the cost.

stuckkeys · 3 years ago
Sure bud. Sure. To each their own. When these GPU enter the consumer market is when the AWS cost becomes irrelevant. =)
pavelstoev · 3 years ago
Often you are better off running certain workloads on lesser GPUs. But this requires certain tricky compiler-level optimizations. For example, can run certain LLM inference with comparable latency on cheaper A40s vs running on A100s. Could also run on 3090s (sometimes even faster). This helps with operating costs but may also resolve availability constraints.
sorcer · 3 years ago
A40 / A6000 and A5000 is a great GPU for single GPU inference and training, provides better price/performance than A100 for models that fit.
1024core · 3 years ago
Doesn't anybody use TPUs from Google?

Given the heterogeneous nature of GPUs, RAM, tensor cores, etc. it would be nice to have a direct comparison of, say, number of teraflops-hour per dollar, or something like that.

TradingPlaces · 3 years ago
Lambda has a very interesting benchmark page https://lambdalabs.com/gpu-benchmarks

If you look through the throughput/$ metric, the V100 16GB looks like a great deal, followed by H100 80GB PCIe 5. For most benchmarks, the A100 looks worse in comparison

dkobran · 3 years ago
FWIW, Paperspace has a similar GPU comparison guide located here https://www.paperspace.com/gpu-cloud-comparison

Disclosure: I work on Paperspace

sva_ · 3 years ago
It seems very honest of you to release this list, considering you're at the very bottom of it by $/hour.
akiselev · 3 years ago
Demand outstrips supply by such a wide margin it doesn’t matter one bit how they compare.

The cheapest 8x A100 (80GB) on the list is LambdaLabs @ $12/hour on demand, and I’ve only once seen any capacity become available in three months of using it. AWS last I checked was $40/hr on demand or $25/hr with 1 year reserve, which costs more than a whole 8xA100 hyperplane from Lambda.

The pricing on these things is nuts right now

slavik81 · 3 years ago
I was very hopeful when I saw "AMD Support" listed on a few of those providers, but that appears to only refer only to AMD CPUs. It is, unfortunately, very difficult to find public cloud providers for AMD GPUs.
pavelstoev · 3 years ago
Very nice. Can we chat briefly ?
KeplerBoy · 3 years ago
why does no one rent out AMD GPUs?

I know those cards are second class citizens in the world of deep learning, but they have had (experimental) pytorch support for a while now, where are the offerings?

mkaic · 3 years ago
> I know those cards are second class citizens in the world of deep learning,

It's worse than that. AMD cards aren't second class citizens, they're not even on the same playing field. ROCm can't compete with CUDA and its ecosystem at all, the most popular deep learning frameworks are only experimentally supported, and Nvidia ships more dedicated tensor processing cores for AI acceleration on their cards. Nvidia has a near monopoly in AI not because they're particularly amazing, but because it seems like AMD is just uninterested in competing.

arvinsim · 3 years ago
I don't think that AMD is uninterested in competing.

It is just that the mindshare is swallowed up by Nvidia that it is really difficult to use something else even if you want to.

detaro · 3 years ago
Why would anyone offer a strictly worse product unless it were lots cheaper to offer, which it isn't? (Even for non-AI use cases, I don't think AMD has much that's more attractive in servers?)
xbmcuser · 3 years ago
They are simply put not price competitive in instruction per dollar at the high end though they are starting to catch up. But to me the biggest reason is software wise they are behind Nvidia. NVidia might be considered a hardware company because of its gpus but they are underappreciated as a the software company that build tools for other to utilise its gpus.
kkielhofner · 3 years ago
They’re not price competitive anywhere in the line.

By the time you spend hours/days/weeks constantly dealing with random edge cases and issues with the poor software support of AMD you could have bought 2-3x the Nvidia hardware (minimum) and still come out ahead.