What can you tell us about this:
> Our internal models (based on Fuyu) have extra capabilities related to our product. In particular,
> 1. They can reliably perform OCR on high-resolution images
> 2. They can do fine-grained localization of text and UI elements within those images
> 3. They can answer questions about images of UIs
Is this just a matter of additional fine tuning, or are there architectural differences?
I'm concerned about the current download's availability - its two URLs to some object storage. I find that these go dark rather quickly for many different reasons (accidentally moving it, bandwidth limits, deleting it later, etc).
I'm curious if there's a reason it's not also hosted on huggingface? I'm not saying they're the best place, but redundancy is good, most models have entries there, they have a very good cdn, and isn't as likely to go dark accidentally.
[1] https://www.adept.ai/