I don't bother to read most Google papers unless someone tells me that they're doing something astounding. Just because I know I don't have access to their models, their code or their data. So what's the point?
As a community we need to stop accepting and stop citing papers like these.
There is no science without replicability, and it is literally impossible to replicate this work. It's not worth the paper it's printed on.
It's fine if Google wants to play with its toys at home. But we should stop pretending this is research of any value.
If it's supposed to stay secret, what's the point of "here's instructions for how to reproduce our big secret"?
Presumably the societal purpose of papers is to share knowledge, and the individual purpose is to take credit and win prestige.
It seems like the first purpose would be better served by also publishing code etc, and the second purpose wouldn't be harmed by it?
Look at GPT-3+, OpenAI gets fame and fortune while people struggle to reproduce their last-gen models.
Then again maybe they’re just using AI as an excuse to get rid of these jobs? Kind of like they did with layoffs only to them outsource many of their departments to other countries for cost savings?
https://arxiv.org/abs/2303.12712
While there is no scientific evidence that LLMs can reach AGI, they will still be practically useful for many other tasks. A human mind paired with an LLM is a powerful combination.
Here’s the thing: the authors of that paper got early access to GPT-4 and ran a bunch of tests on it. The important bit is that MSR does not see into OpenAI’s sausage making.
Now imagine if you were a peasant from 1000 AD who was given a car or TV to examine. Could you really be confident you understood how it worked by just running experiments on it as a black box? If you give a non-programmer the linux kernel, will he/she think it’s magical?
Things look like magic especially when you can’t look under the hood. The story of the Mechanical Turk is one example of that.
Google really killed TF with the transition to TF2. Backwards incompatible everything? This only makes sense if you live in a giant monorepo with tools that rewrite everybody's code whenever you change an interface. (e.g. inside google). On the outside it took TF's biggest asset and turned it into a liability. Every library, blog post, stackoverflow post, etc talking about TF was now wrong. So anybody trying to figure out how to get started or build something was forced into confusion. Not sure about this, but I suspect it's Chollet's fault.
Also, tensorflow was a total nightmare to install while Pytorch was pretty straightforward, which definitely shouldn't be discounted.