Deleted Comment
Deleted Comment
Deleted Comment
You could also look at Vercel and Heroku, I think they do similar things, but I've not used them a lot.
I'd recommend using some boilerplate from github in your favorite stack even if it might be a little bit longer to build a product with it.
> their vision is, at best, like that of a person with myopia seeing fine details as blurry
is a crazy thing to write in an abstract. Did they try to probe that hypothesis at all? I could (well actually I can't) share some examples from my job of GPT-4v doing some pretty difficult fine-grained visual tasks that invalidate this.
Personally, I rate this paper [1], which makes the argument that these huge GenAI models are pretty good at things - assuming that it has seen a LOT of that type of data during training (which is true of a great many things). If you make up tasks like this, then yes can be REALLY bad at them, and initial impressions of AGI get harder to justify. But in practice, we aren't just making up tasks to trip up these models. They can be very performant on some tasks and the authors have not presented any real evidence about these two modes.
It's not that far from reality, most models sees images in very low resolution/limited colors, so not so far from this description
Regular captchas are easily solvable by multimodal LLMs, we're reaching a point where what's hard for software to solve is also hard for humans.
At some point they'll probably have to charge by usage instead of a flat subscription.
The trickier part is “everything else” to make the extension work.