Today, OpenAI showed off its latest large language model (LLM) called GPT-4o — that’s a lower-case “o,” for “omni” — that the company promises can “reason across audio, vision, and text in real time.” During its brief announcement, the company demonstrated the AI’s uncanny ability to assess what it “sees” through the user’s smartphone camera, allowing it to help solve math problems and even assist with coding. OpenAI is making the new model “available to all ChatGPT users, including on the free plan” per OpenAI CEO Sam Altman. “So far, GPT-4 class models have only been available to people who pay a monthly subscription.” It’s arguably a natural evolution of the popular AI chatbot; by harnessing a live video stream, the assistant could likely be more helpful by benefiting from far more context. It’s also unsurprising, considering we’ve seen very similar demos by AI hardware companies Humane and Rabbit, who both attempted to bring an AI chatbot-based gadget with a built-in camera to market this year — albeit with catastrophic results. Given its positioning at the forefront of the tech, however, OpenAI is leveraging the computing power of the modern smartphone instead, and from what we’ve seen, that approach makes for a far more seamless experience, with barely any delays between a user’s question and GPT-4o’s answer. and with video mode!! pic.twitter.com/cpjKokEGVd — Sam Altman (@sama) May 13, 2024 OpenAI claims GPT-40 can respond to audio inputs in as little as 232 milliseconds, which is “similar to human response time in…New ChatGPT AI Watches Through Your Camera, Offers Advice on What It Sees