Readit News logoReadit News
mikesurowiec · 9 months ago
A rough idea of the price differences...

  Per 1k tokens        Input   |  Output
  Amazon Nova Micro: $0.000035 | $0.00014
  Amazon Nova Lite:  $0.00006  | $0.00024
  Amazon Nova Pro:   $0.0008   | $0.0032

  Claude 3.5 Sonnet: $0.003    | $0.015
  Claude 3.5 Haiku:  $0.0008   | $0.0004
  Claude 3 Opus:     $0.015    | $0.075
Source: AWS Bedrock Pricing https://aws.amazon.com/bedrock/pricing/

sheepscreek · 9 months ago
It’s fascinating that Amazon is investing heavily in Anthropic while simultaneously competing with them.
panabee · 9 months ago
Amazon is a retailer and strives to offer choice, whether of books or compute services.

AWS is the golden goose. If Amazon doesn't tie up Anthropic, AWS customers who need a SOTA LLM will spend on Azure or GCP.

Think of Anthropic as the "premium" brand -- say, the Duracell of LLMs.

Nova is Amazon's march toward a house brand, Amazon Basics if you will, that minimizes the need for Duracell and slashes cost for customers.

Not to mention the potential benefits of improving Alexa, which has inexcusably languished despite popularizing AI services.

:Edited for readability

dotBen · 9 months ago
It’s fascinating that Amazon Web Services have so many overlapping and competing services to achieve the same objective. Efficiency/small footprint was never their approach :D

For example, look how many different types of database they offer (many achieve the same objective but different instantiation)

https://aws.amazon.com/products/?aws-products-all.sort-by=it...

donavanm · 9 months ago
As others said the product isnt the model, its the API based token usage. Happily selling whatever model you need, with easy integrations from the rest of your aws stack, is the entire point.
holub008 · 9 months ago
Has anyone found TPM/RPM limits on Nova? Either they aren't limited, or the quotas haven't been published yet: https://docs.aws.amazon.com/general/latest/gr/bedrock.html#l...
tmpz22 · 9 months ago
Maybe they want to gauge demand for a bit first?
Tepix · 9 months ago
I suggest you give the price per million token as seems to be the standard.
8n4vidtmkvmk · 9 months ago
From my personal table https://i.imgur.com/WwL9XkG.png

Price is pretty good. I'm assuming 3.72 chars/tok on average though.. couldn't find that # anywhere.

oblio · 9 months ago
I'm guessing they just copy pasted from the official docs page.
warkdarrior · 9 months ago
Eyeballing it, Nova seems to be 1.5 order of magnitude cheaper than Claude, at all model sizes.

Deleted Comment

Bilal_io · 9 months ago
You have added another zero for Haiku, its output cost is $0.004
indigodaddy · 9 months ago
Thanks that had confused me when I compared same to Nova Pro
mikesurowiec · 9 months ago
You're absolutely right, apologies!
Havoc · 9 months ago
Doesn’t look particularly favourable versus deepseek and qwen. Main deepseek is about same price as smallest nova.

I guess it depends on how sensitive your data is

jerrygoyal · 9 months ago
does anyone know performance benchmark
lukev · 9 months ago
This is a digression, but I really wish Amazon would be more normal in their product descriptions.

Amazon is rapidly developing its own jargon such that you need to understand how Amazon talks about things (and its existing product lineup) before you can understand half of what they're saying about a new thing. The way they describe their products seems almost designed to obfuscate what they really do.

Every time they introduce something new, you have to click through several pages of announcements and docs just to ascertain what something actually is (an API, a new type of compute platform, a managed SaaS product?)

kvakvs · 9 months ago
Amazontalk: We will save you costs Human language: We will make profit while you think you're saving the costs

Amazontalk: You can build on <product name> to analyze complex documents... Human language: There is no product, just some DIY tools.

Amazontalk: Provides the intelligence and flexibility Human language: We will charge your credit card in multiple obscure ways, and we'll be smart about it

Miraste · 9 months ago
That may be generally true, but the linked page says Nova is a series of foundation models in the first sentence.
lukev · 9 months ago
Yeah but even then they won't describe it using the same sort of language that everyone else developing these things does. How many parameters? What kind of corpus was it trained on? MoE, single model, or something else? Will the weights be available?

It doesn't even use the words "LLM", "multimodal" or "transformer" which are clearly the most relevant terms here... "foundation model" isn't wrong but it's also the most abstract way to describe it.

oblio · 9 months ago
Once upon a time there were (and still are) mainframes (and SAP is similar in this respect). These insular systems came with their own tools, their own ecosystem, their own terminology, their own certifications, etc. And you could rent compute & co on them.

If you think of clouds as being cross continent mainframes, a lot more things make a more sense.

danielmarkbruce · 9 months ago
"distributed mainframes".
foobarian · 9 months ago
If you figure out what a security group is, let me know :-D
rsrsrs86 · 9 months ago
Lol

What’s the subnet of the security group of my user group for Aws lambda application in a specific environment that calls kms to get a secret for….

Deleted Comment

jmward01 · 9 months ago
No audio support: The models are currently trained to process and understand video content solely based on the visual information in the video. They do not possess the capability to analyze or comprehend any audio components that are present in the video.

This is blowing my mind. gemini-1.5-flash accidentally knows how to transcribe amazingly well but it is -very- hard to figure out how to use it well and now Amazon comes out with a gemini flash like model and it explicitly ignores audio. It is so clear that multi-modal audio would be easy for these models but it is like they are purposefully holding back releasing it/supporting it. This has to be a strategic decision to not attach audio. Probably because the margins on ASR are too high to strip with a cheap LLM. I can only hope Meta will drop a mult-modal audio model to force this soon.

xendo · 9 months ago
They also announced speech to speech and any to any models for early next year. I think you are underestimating the effort required to release 5 competitive models at the same time.
plumeria · 9 months ago
Is Gemini better than Whisper for transcribing?
jmward01 · 9 months ago
'better' is always a loaded term with ASR. Gemini 1.5 flash can transcribe for 0.01/hour of audio and gives strong results. If you want timing and speaker info you need to use the previous version and a -lot- of tweaking of the prompt or else it will hallucinate the timing info. Give it a try. It may be a lot better for your use case.
ndr_ · 9 months ago
Setting up AWS so you can try it via Amazon Bedrock API is a hassle, so I made a step-by-step guide: https://ndurner.github.io/amazon-nova. It's 14+ steps!
fumeux_fume · 9 months ago
This appears to be a way to steal and harvest aws credentials. No one should be following any of these steps.
ndr_ · 9 months ago
Do you have any evidence for this accusation?

This is a guide for the casual observer who wants to try things out, given that getting started with other AI platforms is so much more straightforward. It's all open source, with transparent hosting, catering to any remaining concerns someone interested in exactly that may have.

metzpapa · 8 months ago
You can set up AWS in less than 5 minutes...
OJFord · 9 months ago
Your 14 steps appear to be 'create an IAM user'..?
Spivak · 9 months ago
If you're already in the AWS ecosystem or have worked in it, it's no problem. If you're used to "make OpenAI account, add credit card, copy/paste API key" it can be a bit daunting.
weitendorf · 9 months ago
AWS does not use the exact same authn/authz/identity model or terminology as other providers, and for people familiar with other models, it's pretty non-trivial to adapt to. I recently posted a rant about this to https://www.reddit.com/r/aws/comments/1geczoz/the_aws_iam_id...

Personally I am more familiar with directly using API keys or auth tokens than AWS's IAM users (which are more similar to what I'd call "service accounts").

scosman · 9 months ago
find to supported region, request model access, wait for model access, create policy, create user, attach policy... it's not comparable
d4rkp4ttern · 9 months ago
Setting up Azure LLM access is a similar hellish process. I learned after several days that I had to look at the actual endpoint URL to determine how to set the “deployment name” and “version” etc.
teruakohatu · 9 months ago
Thanks for that. Are there any proxies that can communicate with bedrock and serve it via a OpenAI style api?
moduspol · 9 months ago
You'd have to deploy it yourself, but there's this:

https://github.com/aws-samples/bedrock-access-gateway

popinman322 · 9 months ago
Try LiteLLM; their core LLM proxy is open source. As an added bonus it also supports other major providers.
SaggyDoomSr · 9 months ago
Nice! FWIW, The only nova model I see on the HuggingFace user space page is us.amazon.nova-pro-v1:0. I cloned the repo and added the other nova options in my clone, but you might want to add them to yours. (I would do a PR, but... I'm lazy and it's a trivial PR :-)).
ndr_ · 9 months ago
OK! I only add what people are interested in, so noted with thanks - will do! :-)
tootie · 9 months ago
I'm so confused on the value prop of Bedrock. It's seems like it wants to be guardrails for implementing RAG with popular models but it's not the least but intuitive. Is it actually better than setting up a custom pipeline?
ndr_ · 9 months ago
The value I get is: 1) one platform, largely one API, several models, 2) includes Claude 3.5 "unlimited" pay-as-you-go, 3) part of our corporate infra (SSO, billing, ... corporate discussions are easier to have)

I'm using none to very little of the functionality they have added recently: not interested in RAG, not interested in Guardrails. Just Claude access, basically.

simonw · 9 months ago
Thank you!
scbenet · 9 months ago
kajecounterhack · 9 months ago
TL;DR comparison of models vs frontier models on public benchmarks here https://imgur.com/a/CKMIhmm
SparkyMcUnicorn · 9 months ago
This doesn't include all the benchmarks.

The one that really stands out is GroundUI-1K, where it beats the competition by 46%.

Nova Pro looks like it could be a SOTA-comparable model at a lower price point.

brokensegue · 9 months ago
So looks like they are trying to win on speed over raw metric performance
zapnuk · 9 months ago
They missed a big opportunity by not offering eu-hosted versions.

Thats a big thing for complience. All LLM-providers reserve the right to save (up to 30days) and inspect/check prompts for their own complience.

However, this means that company data is potentionally sotred out-of-cloud. This is already problematic, even more so when the storage location is outside the EU.

Tepix · 9 months ago
I'm not sure if hosting it in the EU will do any good for Amazon, there's still the US CLOUD Act: It doesn't really matter where the data is located.
physicsguy · 9 months ago
It makes a really big difference for anyone doing business in Europe though.

Legally we're only allowed to use text-embeddings-3-large at work because Azure don't host text-embeddings-3-small within a European region.

ygouzerh · 9 months ago
They might not have enough GPUs datacenters in Europe
xnx · 9 months ago
More options/competition is good. When will we see it on https://lmarena.ai/ ?
glomgril · 9 months ago
looks like it's there now
zacharycohn · 9 months ago
I really wish they would left-justify instead of center-justify the pricing information so I'm not sitting here counting zeroes and trying to figure out how they all line up.