The prerecorded videos, the staff-presented demos, the empty hardware chassis, the suggestive technical details, etc
They have “reasons” for not giving away details, but there are good odds that the ultimate reason is that this is a superficial product update with a lot of flashy patchwork rather than that fundamental advance in AI technology we’d assume from the name.
You can use the product today, right now.
A mining company protecting access to the gold mine is different than a company with a fools gold mine limiting access to the mine to delay analysis.
There might be an "empty chassis" in the academic paper but that's different than tech companies betting on their closed sourced licensing/marketing to spin something less-than-whole.
I wouldn't be suprised if they get millions of new subscribers today.
Here they've managed 2 - it's not open and it's not AI.
How big is this model and what did they do differently (ELI5 please)?
It's good when AI labs don't publish some details about powerful models, for the same reason that it's good when bio research labs don't publish details about dangerous viruses.
To your second question, I am worried about the power dynamics of one lab having a monopoly on super-powerful models. But by far the worst risk I'm worried about (and it's my job to try and help mitigate) is catastrophic accidents from someone creating a super-powerful model without the right alignment techniques and safeguards. And that kind of risk is heightened when there are more actors competitively racing to build AGI.
I can't tell if the answers of the tool are objective or if the answers aren't biases by the training corpus (or some collective bias ingrained within the training corpus). I really can't tell much about it. It's very much a black box - take it or leave it.
I think that's a great deal.
I mean science is all about objectivity (or so i was told)