I have an intuition about Suleyman, that being that his marketing background make him an incredibly unreliable source of actual information. He makes a lot of big predictions on future AI capabilities, like for hallucinations as a recent example I can think of and engages heavily in hype drumming in his interviews and social media. The untrustworthy aura I feel around the company extends to their products. Inflection-1′s technical paper (I can’t find a potential longer version) is very short compared to GPT-4 or PALM-2 and is entirely pictures of condensed benchmark results with a few paragraphs of explanations.
I expect my views to be wrong, but for now while inflection definitely has the compute, I have a feeling there’s a lot more limits and smoke involved that wouldn’t quite put them up with OpenAI, DeepMind and Meta in terms of impact on both the market and AGI timelines.
Edit: Originally misread the computer calculations from the post and used my mistake as evidence of my first point. Though the evidence is no longer there, it was confirming an intuition I already had and still stand by.
I’m not sure what you’re referring to. The “three times the compute for GPT-4” claim places their compute at 6.7e17, which is lower than the estimates from the other two methods.
I have an intuition about Suleyman, that being that his marketing background make him an incredibly unreliable source of actual information. He makes a lot of big predictions on future AI capabilities, like for hallucinations as a recent example I can think of and engages heavily in hype drumming in his interviews and social media. The untrustworthy aura I feel around the company extends to their products. Inflection-1′s technical paper (I can’t find a potential longer version) is very short compared to GPT-4 or PALM-2 and is entirely pictures of condensed benchmark results with a few paragraphs of explanations.
I expect my views to be wrong, but for now while inflection definitely has the compute, I have a feeling there’s a lot more limits and smoke involved that wouldn’t quite put them up with OpenAI, DeepMind and Meta in terms of impact on both the market and AGI timelines.
Edit: Originally misread the computer calculations from the post and used my mistake as evidence of my first point. Though the evidence is no longer there, it was confirming an intuition I already had and still stand by.
I’m not sure what you’re referring to. The “three times the compute for GPT-4” claim places their compute at 6.7e17, which is lower than the estimates from the other two methods.
You’re right, I completely misread it. I’ll edit my comment with that in mind.