It's going to be a hot, hot summer. When they talk about us moving into an accelerating phase, they aren't kidding:
OpenAI, Google, and Mistral have released new AI models within 12 hours. The fact that both GPT-4 Turbo and Gemini Pro 1.5 are 'multimodal' systems will open up many new use cases; now, we can use images, audio, and video in our prompts.
I recently listened to the Lex Fridman Podcast with Yann LeCun (Meta's Chief AI Scientist). He's quite unimpressed by the current LLMs' ability to understand the world. So, his take on researchers focusing on 'objective-driven' AI with the ability to reason and plan to bring about better world understanding is intriguing.
He uses the colorful metaphor of a child locked in a closed, windowless room with a giant pile of books who is somehow expected to make sense of the world. The reality is that children need to run outside and see things with their own eyes, play with sticks, climb trees, and so on. You learn about the world by seeing and doing, not just reading.
Embodied AI (ubiquitous robots) are going to be a huge boost here.
#openai #google #meta #mistral #robotics #ai #llm
---
“We hear a lot of people saying: ‘Oh my God, we’re going to get [artificial general intelligence] within the next year,” said Meta’s chief AI scientist, Yann LeCun, responding to a claim from the xAI founder, Elon Musk. “It’s just not happening. We have AI systems that can pass the bar exam, but they can’t clear up your dinner table and fill up the dishwasher. We have systems that manipulate language, and fool us into thinking that they are smart, but cannot understand the world.”