Readit News logoReadit News
mikeocool · a year ago
Curious if anyone knows the logistics of these cloud provider/AI company deals. In this case, it seems like the terms of the deal mean that Anthropic ends up spending most of the investment on AWS to pay for training.

Does anthropic basically get at cost pricing on AWS? If Amazon has any margin on their pricing, it seems like this $4B investment ends up costing them a lot less, and this is a nice way to turn a cap ex investment into AWS revenue.

tyre · a year ago
Yes exactly.

This was the brilliance of the original MSFT investment into OpenAI. It was an investment in Azure scaling its AI training infra, but roundabout through a massive customer (exactly what you’d want as a design partner) and getting equity.

I’m sure Anthropic negotiated a great deal on their largest cost center, while Amazon gets a huge customer to build out their system with.

wcunning · a year ago
That’s honestly one of the hardest things in engineering — identifying not just a customer to drive requirements, but a knowledgeable customer who can drive good requirements that work for a broader user base and can drive further expansion. Anthropic seems ideal for that, plus they act as a service/API provider on AWS.
dzonga · a year ago
or simply one of the best corporate buyouts that's not subject to regulatory scrutiny. microsoft owns 49% of OpenAI - will get profits till whenever. All without subject to regulatory approval. and they get to improve Azure
fny · a year ago
And Amazon can always build their own LLM product down the line. Building out data centers feels like a far more difficult problem.
whatshisface · a year ago
This explanation makes no sense, I could be AWS' biggest customer if they wanted to pay me for it. Something a little closer could be that the big tech companies wanted to acquire outside LLMs, not quite realizing that spending $1B on training only puts you $1B ahead.
eitally · a year ago
I am not privy to specific details, but in general there is a difference between investment and partnership. If it's literally an investment, it can either be in cash or in kind, where in kind can be like what MSFT did for OpenAI, essentially giving them unlimited-ish ($10b) Azure credits for training ... but there was quid pro quo where MSFT in turn agreed to embed/extend OpenAI in Azure services.

If it's a partnership investment, there may be both money & in-kind components, but the money won't be in the context of fractional ownership. Rather it would be partner development funds of various flavors, which are usually tied to consumption commits as well as GTM targets.

Sometimes in reading press releases or third party articles it's difficult to determine exactly what kind of relationship the ISV has with the CSP.

B4CKlash · a year ago
There's also another angle. During the call with Lex last week, Dario seemed to imply that future models would run on amazon chips from Annapurna Labs (Amazon's 2015 fabless purchase). Amazon is all about the flywheel + picks and shovels and I, personally, see this as the endgame. Create demand for your hardware to reduce the per unit cost and speed up the dev cycle. Add the AWS interplay and it's a money printing machine.
shawndrost · a year ago
You can find the text of the original OpenAI/MSFT deal here: https://www.lesswrong.com/posts/5jjk4CDnj9tA7ugxr/openai-ema...
scosman · a year ago
Also: they need top tier models for their Bedrock business. They are one of only a few providers for Claude 3.5 - it’s not open and anthropic doesn’t let many folks run it.

Google has Gemini (and Claude), MSFT has OpenAI. Amazon needs this to stay relevant.

chatmasta · a year ago
Supermicro is currently under DOJ investigation for similar schemes to this. The legality of it probably depends on the accounting, and how revenue is recognized, etc.

It certainly looks sketchy. But I’m sure there’s a way to do it legitimately if their accountants and lawyers are careful about it…

DAGdug · a year ago
This assumes they have no constraint when it comes to supply, and therefore no opportunity cost.
dustingetz · a year ago
i believe they get to book any investment of cloud credits as revenue, here’s a good thread explaining the grift: https://news.ycombinator.com/item?id=39456140 basically you’re investing your own money in yourself which mostly nets out but you get to keep the equity (and then all the fool investors FOMO in on fake self dealing valuations)
yard2010 · a year ago
...isn't it tax fraud with extra steps? Asking seriously.
aiinnyc · a year ago
One hand washes the other.
paulddraper · a year ago
Correct.

Same with Microsoft.

peppertree · a year ago
Anthropic should double down on the strategy of being the better code generator. No I don't need an AI agent to call the restaurant for me. Win the developers over and the rest will follow.
rtsil · a year ago
> Win the developers over and the rest will follow.

Will they really? Anecdotal evidence, but nobody I know in real life knows about Claude (other than it's an ordinary first name). And they all use or at least know about ChatGPT. None of them are software engineers of course. But the corporate deciders aren't software engineers either.

999900000999 · a year ago
Normal people aren't paying for LLMs.

If they ever do Apple and Google will offer it as a service built into your phone .

For example, you could say ok Google call that restaurant me and My girlfriend had our first date at 5 years ago, set up something nice so I can propose. And I guess Google Gemini ( or whatever it's called at this point), Will hire a band, some photographers, and maybe even a therapist just in case it doesn't work out.

All of this will be done seamlessly.

But I don't imagine any normal person will pay 20 or $30 a month for a standalone service doing this. As is it's going to be really hard to compete against GitHub Copilot they effectively block others from scrapping GitHub.

peppertree · a year ago
Consumers don't have to consciously choose Claude, just like most people don't know about Linux. But if they use an Android phone or ever use any web services they are using Linux.
findjashua · a year ago
Every single person I know who pays for an LLM is a developer who pays for Claude because of coding ability
staticman2 · a year ago
Most people I know in real life have certainly heard of ChatGPT but don't pay for it.

I think someone enthusiastic enough to pay for the subscription is more likely to be willing to try a rival service, but that's not most people.

Usually when these services are ready to grow they offer a month or more free to try, at least that's what Google has been doing with their Gemini bundle.

ramraj07 · a year ago
OP and the people who reply to you are perfect examples of engineers being clueless about how the rest of the world operates. I know engineers who don’t know Claude, and I know many, many regular folk who pay for ChatGPT (basically anyone who’s smart and has money pays for it). And yet the engineers think they understand the world when in reality they just understand how they themselves work best.
ToDougie · a year ago
I use Claude Pro paid version every day, but not for coding. I used to be a software engineer, but no longer. I tried OpenAI in the past, but I did not enjoy it. I do not like Sam Altman.

My use cases: Generating a business plan, podcast content, marketing strategies, sales scripts, financial analyses, canned responses, and project plans. I also use it for general brainstorming, legal document review, and so many other things. It really feels like a super-assistant.

Claude has been spectacular about 98% of the time. Every so often it will refuse to perform an action - most recently it was helping me research LLC and trademark registrations, combined with social media handles (and some deviations) and web URL availability. It would generate spectacular reports that would have taken me hours to research, in minutes. And then Claude decided that it couldn't do that sort of thing, until it could the next day. Very strange.

I have given Gemini (free), OpenAI (free and Paid), Copilot (free), Perplexity (free) a shot, and I keep coming back to Claude. Actually, Copilot was a pretty decent experience, but felt the guardrails too often. I do like that Microsoft gives access to Dall-E image generation at no cost (or maybe it is "free" with my O365 account?). That has been helpful in creating simple logo concepts and wireframes.

I run into AI with Atlassian on the daily, but it sucks. Their Confluence AI tool is absolute garbage and needs to be put down. I've tried AI tools that Wix, Squarespace, and Mira provide. Those were all semi-decent experiences. And I just paid for X Premium so I can give Grok a shot. My friend really likes it, but I don't love the idea of having to open an ultra-distracting app to access it.

I'm hoping some day to be like the wizards on here who connect AI to all sorts of "things" in their workflows. Maybe I need to learn how to use something like Zapier? If I have to use OpenAI with Zapier, I will.

If you read this far, thanks.

teaearlgraycold · a year ago
They’ll use whatever LLM is integrated into the back end of their apps. And the developers have the most sway over that.
croes · a year ago
Maybe the software engineers should talk to the deciders then.
bambax · a year ago
In my experience*, for coding, Sonnet is miles above any model by OpenAI, as well as Gemini. They're all far from perfect, but Sonnet actually "gets" what you're asking, and tries to help when it fails, while the others wander around and often produce dismal code.

* Said experience is mostly via OpenRouter, so it may not reflect the absolute latest developments of the models. But there at least, the difference is huge.

fullstackwife · a year ago
I also don't understand the idea of voice mode, or agent controller computer. Maybe it is cool to see as a tech demo, but all I really want is good quality, at reasonable price for the LLM service
lxgr · a year ago
I think voice mode makes significantly more sense when you consider people commuting by car by themselves every day.

Personally I don't (and I'd never talk to an LLM on public transit or in the office), but almost every time I do drive somewhere, I find myself wishing for a smarter voice-controlled assistant that would allow me to achieve some goal or just look up some trivia without ever having to look at a screen (phone or otherwise).

wenc · a year ago
Voice mode can be useful when you're reading a (typically non-fiction) book and need to ask the LLM to check something.

It's essentially a hands-free assistant.

YZF · a year ago
Developers, developers, developers!

More seriously: I think there are a ton of potential applications. I'm not sure that developers that use AI tools are more likely to build other AI products - maybe.

yard2010 · a year ago
Reference for the memberberries: https://youtu.be/Vhh_GeBPOhs
zamderax · a year ago
No they should not do this. They are trying to create generalized artificial intelligence not a specific one. Let the cursor, zed, codeium or some smaller company focus on that.
rty32 · a year ago
I wonder at OpenAI, Anthropic etc, how many people actually believe in "creating generalized artificial intelligence"
paxys · a year ago
Which use case do you think benefits more regular customers around the world?
hehehheh · a year ago
Which use case generates more revenue? (Genuine question. It could be the restaurants but how to monitize)
ianmcgowan · a year ago
I mean, look at Linux and Firefox!
gopalv · a year ago
> look at Linux and Firefox!

AI models are more like a programming language or CPU architecture.

OpenAI is Intel and Anthropic is AMD.

peppertree · a year ago
Pretty sure most frontend developers use Chrome since it has better dev tools. And yes everyone uses Linux most just don't know it.
ripped_britches · a year ago
Legendary comment, bravo
cainxinth · a year ago
They certainly need the money. The Pro service has been running in limited mode all week due to being over capacity. It defaults to “concise” mode during high capacity but Pro users can select to put it back into “Full Response.” But I can tell the quality drops even when you do that, and it fails and brings up error messages more commonly. They don’t have enough compute to go around.
jmathai · a year ago
I’ve been using the API for a few weeks and routinely get 529 overloaded messages. I wasn’t sure if that’s always been the case but it certainly makes it unsuitable for production workloads because it will last hours at a time.

Hopefully they can add the capacity needed because it’s a lot better than GPT-4o for my intended use case.

rmbyrro · a year ago
Sonnet is better than 4o for virtually all use cases.

The only reason I still use OpenAI's API and chatbot service is o1-preview. o1 is like magic. Everything Sonnet and 4o do poorly, o1 solves like a piece of cake. Architecting, bug fixing, planning, refactoring, o1 has never let me know on any 'hard' task.

A nice combo is have o1 guiding Sonnet. I ask o1 to come up with a solution and explanation, then simply feed its response into Sonnet to execute. That running on Aider really feels like futuristic stuff.

AlexAndScripts · a year ago
Amazon Bedrock supports Claude 3.5, and you can use inference profiles to split it across multiple regions. It's also the same price.

For my use case I use a hybrid of the two, simulating standard rate limits and doing backoff on 529s. It's pretty reliable that way.

Just beware that the European AWS regions have been overloaded for about a month. I had to switch to the American ones.

shmatt · a year ago
in the beginning i was agitated by Concise and would move it back manually. But then I actually tried it, I asked for SQL and it gave me back SQL and 1-2 sentences at most

Regular mode gives SQL and entire paragraphs before and after it. Not even helpful paragraphs, just rambling about nothing and suggesting what my next prompt should be

Now I love concise mode, it doesn't skimp on the meat, just the fluff. Now my problem is, concise only shows up during load. Right now I can't choose it even if i wanted to

cruffle_duffle · a year ago
Totally agree. I wish there was a similar option on ChatGPT. These things are seemingly trained to absolutely love blathering on.

And all that blathering eats into their precious context window with tons of repetition and little new information.

nmfisher · a year ago
Agree, concise mode is much better for code. I don’t need you to restate the request or summarize what you did. Just give me the damn code.
el_benhameen · a year ago
Interesting. I also find it frustrating to be rate limited/have responses fail when I’m paying for the product, but I’ve actually found that the “concise” mode answers have less fluff and make for faster back and forth. I’ve once or twice looked for the concise mode selector when the load wasn’t high.
rvz · a year ago
All that money and talk of "scale" and yet not only it is slow but costs billions a year to run at normal load and is struggling at high load.

This is essentially Google-level load and they can't do it.

johnisgood · a year ago
Agreed, I was surprised by it after I first have subscribed to Pro and had a not-that-long chat with it.
moffkalast · a year ago
Their shitty UI is also not doing them any infrastructure favors, during load it'll straight up write 90% of an answer, and then suddenly cancel and delete the whole thing, so you have to start over and waste time generating the entire answer again instead of just continuing for a few more sentences. It's like a DDOS attack where everyone gets preempted and immediately starts refreshing.
wis · a year ago
Yes! It's infuriating when Claude stops generating mid response and deletes the whole thread/conversation. Not only you lose what it has generated so far, which would've been at least somewhat useful, but you also lose the prompt you wrote, which could've taken you some effort to write.
cma · a year ago
> But I can tell the quality drops even when you do that

Dario said in a recent interview that they never switch to a lower quality model in terms of something with different parameters during times of load. But he left room for interpretation on whether that means they could still use quantization or sparsity. And then additionally, his answer wasn't clear enough to know whether or not they use a lower depth of beam search or other cheaper sampling techniques.

He said the only time you might get a different model itself is when they are A-B testing just before a new announced release.

And I think he clarified this all applied to the webui and not just the API.

(edit: I'm rate limited on hn, here's the source in reply to the below https://www.youtube.com/watch?v=ugvHCXCOmm4&t=42m19s )

dr_dshiv · a year ago
Rate limited on hn! Share more please
avarun · a year ago
Source?
nowahe · a year ago
I've had it refuse to generate a long text response (I was trying to concise a 300kb documentation to 20-30kb to be able to put it in the project's context), and every time I asked it replied "How should structure the results ?", "Shall I go ahead with writing the artifacts now ?", etc.

It wasn't even during the over-capacity event I don't think, and I'm a pro user.

Filligree · a year ago
Hate to be that guy, but did you tell it up front not to ask? And, of course, in a long-running conversation it's important not to leave such questions in the context.
neya · a year ago
I am a paying customer with credits and the API endpoints rate-limited me to the point where it's actually unusable as a coding assistant. I use a VS Code extension and it just bailed out in the middle of a migration. I had to revert everything it changed and that was not a pleasant experience, sadly.
square_usual · a year ago
When working with AI coding tools commit early, commit often becomes essential advice. I like that aider makes every change its own commit. I can always manicure the commit history later, I'd rather not lose anything when the AI can make destructive changes to code.
teaearlgraycold · a year ago
Why not just continue the migration manually?
htrp · a year ago
Control your own inference endpoints.
datavirtue · a year ago
You aren't running against a local LLM?
0xDEAFBEAD · a year ago
More evidence that people should use wrappers like OpenRouter and litellm by default? (Makes it easy to change your choice of LLMs, if one is experiencing problems)
llm_trw · a year ago
Neither does OAI. Their service has been struggling for more than a week now. I guess everyone is scrambling after the new qwen models dropped and matched the current state of the art with open weights.
sbuttgereit · a year ago
Hmmm... I wonder if this is why some of the results I've gotten over the past few days have been pretty bad. It's easy to dismiss poor results on LLM quality variance from prompt to prompt vs. something like this where the quality is actively degraded without notification. I can't say this is in fact what I'm experience, but it was noticeable enough I'm going to check.
jmathai · a year ago
Never occurred to me that the response changes based on load. I’ve definitely noticed it seems smarter at times. Makes evaluating results nearly impossible.
Seattle3503 · a year ago
This is one reason closed models suck. You can't tell if the bad responses are due to something you are doing, or if the company you are paying to generate the responses is cutting corners and looking for efficiencies, eg by reducing the number of bits. It is a black box.
baxtr · a year ago
Recently I started wondering about the quality of ChatGPT. A couple of instances I was like: "hmm, I’m not impressed at all by this answer, I better google it myself!"

Maybe it’s the same effect over there as well.

55555 · a year ago
Same experience here.

Deleted Comment

demaga · a year ago
I think Claude is actually superior to ChatGPT and needs more recognition. So good news, I guess
internet101010 · a year ago
Yep. I start most technical prompts with 4o and Claude side-by-side in LibreChat and more often than not end up moving forward with Claude.
r0fl · a year ago
I agree it’s better for coding but it hits limits or seems very slow , even on paid subscription, a lot more often than ChatGPT
jatins · a year ago
Anthropic gets a lot of it's business via AWS Bedrock so it's fair to say that Amazon probably has reasonable insight into how the Claude usage is growing that makes them confident in this investment
paxys · a year ago
They are also confident in the investment because they know that all the money is going to come right back to them in the short term (via AWS spending) whether or not Anthropic actually survives in the long term.
VirusNewbie · a year ago
But anthropic is currently on GCP.
swyx · a year ago
> gets a lot of it's business via AWS Bedrock

can you quantify? any numbers, even guesstimates?

mediaman · a year ago
One source [1] puts it at 60-75% of revenue as third-party API, most of which is AWS.

[1]https://www.tanayj.com/p/openai-and-anthropic-revenue-breakd...

apwell23 · a year ago
> Anthropic gets a lot of it's business via AWS Bedrock

How do you know this

ramesh31 · a year ago
Anthropic will be the winner here, zero doubts in my mind. They have leapfrogged head and shoulders above OpenAI over the last year. Who'd have thought a business predicated entirely on keeping the ~1000 people on earth qualified to work on this stuff happy would go downhill once they failed at that.
fariszr · a year ago
This makes sense in the grand scheme of things. Anthropic used to be in the Google camp, but DeepMind seems to have picked up speed lately, with new “Experimental” Gemini Models beating everyone, while AWS doesn't have anything on the cutting edge of AI.

Hopefully this helps Anthropic to fix their abysmal rate limits.

n2d4 · a year ago
> Anthropic used to be in the Google camp

I don't think Anthropic took any allegiances here. Amazon already invested $4B last year (Google invested $2B).

fariszr · a year ago
AFAIK they used Gcloud to run their models.
submeta · a year ago
I had to switch from Pro to Teams plan and pay 150 USD for 5 accounts because the Pro plan has gotten unusable. It will allow me to ask a dozen or so questions and then will block me for hours because of „high capacity.“ I don’t need five accounts, one for 40 USD would be totally fine if it would allow me to work uninterrupted for a couple of hours.

All in all Claude is magic. It feels like having ten assistants at my fingertip. And for that even 100 USD is worth paying.

modriano · a year ago
I just start new chats whenever the chat gets long (in terms of number of tokens). It's kind of a pain to have to form a prompt that encapsulates enough context, but it has prevented me from hitting the Pro limit. Also, I include more questions and detail in each prompt.

Why does that work? Claude includes the entire chat with each new prompt you submit [0], and the limit is based on the number of tokens you've submitted. After not too many prompts, there can be 10k+ tokens in the chat (which are all submitted in each new prompt, quickly advancing towards the limit).

(I also have a chatGPT sub and I use that for many questions, especially now that it includes web search capabilities)

[0] https://support.anthropic.com/en/articles/8324991-about-clau...

greenie_beans · a year ago
> It's kind of a pain to have to form a prompt that encapsulates enough context, but it has prevented me from hitting the Pro limit. Also, I include more questions and detail in each prompt.

i get it to provide a prompt to start the new chat. i sometimes wish there was a button for it bc it's such a big part of my workflow

esperent · a year ago
Why don't you use the API with LibreChat instead?
submeta · a year ago
Can I replicate the „Projects“ feature where I upload files and text to give context? And will it allow me to follow up on previous chats?