Artificial intelligence is more a part of our lives than ever before. While some might call it hype and compare it to NFTs or 3D TVs, AI is causing a sea change in nearly every facet of life that technology touches. Bing wants to know you intimately, Bard wants to reduce websites to easy-to-read cards, and ChatGPT has infiltrated nearly every part of our lives. At The Verge, we’re exploring all the good AI is enabling and all the bad it’s bringing along.
Featured stories
After a few years working on Artifact, Mike Krieger has a new AI project to work on.
At Google I/.O 2024 today, Google announced a multimodal version of Gemini Nano, allowing the on-device processing-powered AI model to recognize images, sounds, and spoken language in addition to text.
Those multimodal capabilities are also coming to the Android accessibility feature TalkBack, using AI to fill in missing information about unlabeled images, without requiring a connection to the internet.
Google even highlighted the wrong answer in the video!
How we use the internet is changing fast thanks to the advancement of AI-powered chatbots that can find information and redeliver it as a simple conversation.
Obviously, someone noticed our video that clipped every single AI mention at I/O 2023 last year. Sundar Pichai closed the 2024 keynote by showing how AI can save us some work by using it to keep track. At the time, it was up to 121 AI mentions.
...by the time they were finished, it was probably more like 124.
Head over to Google’s Vertex AI Studio site and click “Try it in console” to goof around with some of the AI tools Google talked about at I/O today. The site is meant for developers who want to test the company’s models out while deciding what works best for their software, but anyone can play with it.
OpenAI didn’t mean to kickstart a generational shift in the technology industry. But it did. Now all we have to decide is where to go from here.
Google teased the next generation of its small language model Gemma, including a larger version with 27 billion parameters (or how much a model understands).
The company also announced PaliGemma, an open-source model in the Gemma family for labeling photos and adding captions to images.
Google’s Sameer Samat just made very clear how much Google sees Gemini as a way to make Android a better operating system. Gemini app? On iOS? Yeah yeah sure sure. But AI on Android, and particularly on Pixel, is where the real stuff is for Google. The AI race is very much a smartphone race, friends.
Toward the end (maybe?) of the I/O keynote, Google threw in a cute little ditty about all the things you can do with Gemini prompts: generate photos of cats playing guitar, find smart things to say about Renoir, etc.
It includes the phrase “There’s no wrong way to prompt,” which, have you met people?
Autonomous Cars
Waymo’s robotaxis are under investigation for crashes and traffic law violations
Cruise is back driving autonomously for the first time since pedestrian-dragging incident
Amazon’s robotaxi company is under investigation after two crashes with motorcyclists
50,000.
With more AI features coming to Google Workspace and other Google products, customers might be wondering if this means the next Gemini version learned from their emails. Google says it will not use user files on its platform to teach Gemini or other AI models.
When you “go live” — I guess that’s what we’re calling it — you can wave around your smartphone camera and ask about what’s around you in real time. Like OpenAI’s GPT-4o, you can even interrupt it. (It is not clear if it sounds like ScarJo.)
The company’s new Gemini voice chat feature will come out “later this year.”
Google added a new feature to Workspace that lets users ask an AI agent questions about meetings, emails, and everything else that someone may need to know at work. The AI Teammate, renamed Chip during Google I/O, takes information from company data to answer questions.
Here’s a quick look at the new multimodal AI project Google just announced that’s called Astra and how it can help you find misplaced glasses.
Note: this video was edited for length and clarity, but the original video was one single take.
Google invented a lot of core AI technology, and now the company’s turning to Demis to get back in front of the AI race for AI breakthroughs.
Google added AI into Search so people who don’t like planning that much (aka me) can cosplay an organized person. It lets people use Google Search to find the best restaurants for their specific event, create a meal plan with food they actually like, or organize trips with friends.
Onstage at Google I/O, the company is showing off its AI enhanced Google Search that can take a complex and long entry and generate an AI Overview. In an example, Google searched “Find the best yoga or pilates studios in Boston and show details on their intro offers and walking time from Beacon Hill,” and it provided all of that for the user.
1/6