Alongside unveiling its latest foundation model, Gemini, and making big claims about its capabilities, Google also released a demo video on YouTube, where a user can be seen doing various activities on a table, and the AI was able to recognize it instantly and verbally respond. The shocking video was met with awe because no one in the industry has been able to achieve such results from an AI model so far, and likely nobody will for a long time – including Google. It turns out the video was not entirely real and Google edited the Gemini AI demo video and added elements to make it seem more advanced than it really is.
Questions were raised on the demo video first by a Bloomberg op-ed by Parmy Olsen who claimed that Google was misrepresenting the capabilities of its Gemini AI in the demo. Later, The Verge reached out to aGoogle spokesperson who then referred to a post on X by Oriol Vinyals, the vice president of Research & Deep Learning Lead at Google DeepMind. He is also the co-lead for the Gemini project.
In a post, he said, “All the user prompts and outputs in the video are real, shortened for brevity. The video illustrates what the multimodal user experiences built with Gemini could look like. We made it to inspire developers”.