Connect with us

Technologies

The Future’s Here: Testing Out Gemini’s Live Camera Mode

Gemini Live’s new camera mode feels like the future when it works. I put it through a stress test with my offbeat collectibles.

«I just spotted your scissors on the table, right next to the green package of pistachios. Do you see them?»

Gemini Live’s chatty new camera feature was right. My scissors were exactly where it said they were, and all I did was pass my camera in front of them at some point during a 15-minute live session of me giving the AI chatbot a tour of my apartment. Google’s been rolling out the new camera mode to all Android phones using the Gemini app for free after a two-week exclusive to Pixel 9 (including the new Pixel 9A) and Galaxy S5 smartphones. So, what exactly is this camera mode and how does it work?

When you start a live session with Gemini, you now how have the option to enable a live camera view, where you can talk to the chatbot and ask it about anything the camera sees. Not only can it identify objects, but you can also ask questions about them — and it works pretty well for the most part. In addition, you can share your screen with Gemini so it can identify things you surface on your phone’s display. 

When the new camera feature popped up on my phone, I didn’t hesitate to try it out. In one of my longer tests, I turned it on and started walking through my apartment, asking Gemini what it saw. It identified some fruit, ChapStick and a few other everyday items with no problem. I was wowed when it found my scissors. 

That’s because I hadn’t mentioned the scissors at all. Gemini had silently identified them somewhere along the way and then  recalled the location with precision. It felt so much like the future, I had to do further testing. 

My experiment with Gemini Live’s camera feature was following the lead of the demo that Google did last summer when it first showed off these live video AI capabilities. Gemini reminded the person giving the demo where they’d left their glasses, and it seemed too good to be true. But as I discovered, it was very true indeed.

Gemini Live will recognize a whole lot more than household odds and ends. Google says it’ll help you navigate a crowded train station or figure out the filling of a pastry. It can give you deeper information about artwork, like where an object originated and whether it was a limited edition piece.

It’s more than just a souped-up Google Lens. You talk with it, and it talks to you. I didn’t need to speak to Gemini in any particular way — it was as casual as any conversation. Way better than talking with the old Google Assistant that the company is quickly phasing out.

Google also released a new YouTube video for the April 2025 Pixel Drop showcasing the feature, and there’s now a dedicated page on the Google Store for it.

To get started, you can go live with Gemini, enable the camera and start talking. That’s it.

Gemini Live follows on from Google’s Project Astra, first revealed last year as possibly the company’s biggest «we’re in the future» feature, an experimental next step for generative AI capabilities, beyond your simply typing or even speaking prompts into a chatbot like ChatGPT, Claude or Gemini. It comes as AI companies continue to dramatically increase the skills of AI tools, from video generation to raw processing power. Similar to Gemini Live, there’s Apple’s Visual Intelligence, which the iPhone maker released in a beta form late last year. 

My big takeaway is that a feature like Gemini Live has the potential to change how we interact with the world around us, melding our digital and physical worlds together just by holding your camera in front of almost anything.

I put Gemini Live to a real test

The first time I tried it, Gemini was shockingly accurate when I placed a very specific gaming collectible of a stuffed rabbit in my camera’s view. The second time, I showed it to a friend in an art gallery. It identified the tortoise on a cross (don’t ask me) and immediately identified and translated the kanji right next to the tortoise, giving both of us chills and leaving us more than a little creeped out. In a good way, I think.

I got to thinking about how I could stress-test the feature. I tried to screen-record it in action, but it consistently fell apart at that task. And what if I went off the beaten path with it? I’m a huge fan of the horror genre — movies, TV shows, video games — and have countless collectibles, trinkets and what have you. How well would it do with more obscure stuff — like my horror-themed collectibles?

First, let me say that Gemini can be both absolutely incredible and ridiculously frustrating in the same round of questions. I had roughly 11 objects that I was asking Gemini to identify, and it would sometimes get worse the longer the live session ran, so I had to limit sessions to only one or two objects. My guess is that Gemini attempted to use contextual information from previously identified objects to guess new objects put in front of it, which sort of makes sense, but ultimately, neither I nor it benefited from this.

Sometimes, Gemini was just on point, easily landing the correct answers with no fuss or confusion, but this tended to happen with more recent or popular objects. For example, I was surprised when it immediately guessed one of my test objects was not only from Destiny 2, but was a limited edition from a seasonal event from last year. 

At other times, Gemini would be way off the mark, and I would need to give it more hints to get into the ballpark of the right answer. And sometimes, it seemed as though Gemini was taking context from my previous live sessions to come up with answers, identifying multiple objects as coming from Silent Hill when they were not. I have a display case dedicated to the game series, so I could see why it would want to dip into that territory quickly.

Gemini can get full-on bugged out at times. On more than one occasion, Gemini misidentified one of the items as a made-up character from the unreleased Silent Hill: f game, clearly merging pieces of different titles into something that never was. The other consistent bug I experienced was when Gemini would produce an incorrect answer, and I would correct it and hint closer at the answer — or straight up give it the answer, only to have it repeat the incorrect answer as if it was a new guess. When that happened, I would close the session and start a new one, which wasn’t always helpful.

One trick I found was that some conversations did better than others. If I scrolled through my Gemini conversation list, tapped an old chat that had gotten a specific item correct, and then went live again from that chat, it would be able to identify the items without issue. While that’s not necessarily surprising, it was interesting to see that some conversations worked better than others, even if you used the same language. 

Google didn’t respond to my requests for more information on how Gemini Live works.

I wanted Gemini to successfully answer my sometimes highly specific questions, so I provided plenty of hints to get there. The nudges were often helpful, but not always. Below are a series of objects I tried to get Gemini to identify and provide information about. 

Technologies

OpenAI Launches ChatGPT Atlas, Challenging Google Chrome With an AI-First Browser

The browser is available now for MacOS users, with versions for Windows, iOS and Android coming later.

OpenAI has released a generative AI-powered web browser called ChatGPT Atlas, a major step in the company’s expansion beyond its ChatGPT chatbot platform. The browser, announced Tuesday, integrates ChatGPT’s capabilities directly into the browsing experience, aiming to make web use more interactive and chatbot-like.

OpenAI sparked speculation earlier Tuesday after posting a teaser on its X account showing a series of browser tabs. During the YouTube livestream, CEO Sam Altman and others announced the browser and live-demoed a few of the new features now available for MacOS users worldwide. Support for Windows, iOS and Android operating systems is «coming soon,» the company said.

(Disclosure: Ziff Davis, CNET’s parent company, in April filed a lawsuit against OpenAI, alleging it infringed Ziff Davis copyrights in training and operating its AI systems.)

The new product launch comes amid growing competition among tech companies to embed AI assistants more deeply into everyday tools. For instance, Google has already integrated Gemini into its Chrome browser to add AI to the online browsing experience. Earlier this year, the AI search tool developer Perplexity launched Comet, an AI-powered Chromium-based web browser. Here’s everything OpenAI announced today.


Don’t miss any of our unbiased tech content and lab-based reviews. Add CNET as a preferred Google source.


What is ChatGPT Atlas?

ChatGPT Atlas looks and functions like a traditional web browser. It includes tabs, bookmarks, extensions and incognito mode, but adds popular ChatGPT functions and features throughout. Opening a new tab lets you either enter a URL or ask ChatGPT a question. The browser includes separate tabs for different types of results, such as search links, images, videos and news.

A built-in ChatGPT sidebar can analyze whatever page you’re viewing to provide summaries, explanations or quick answers without leaving the site. ChatGPT can also offer in-line writing assistance, suggesting edits and completions inside any text field, such as an email draft. 

One of the biggest new features is browser memory, which keeps track of pages and topics you’ve previously explored. Atlas can suggest related pages, help you return to past research or automate repetitive tasks. Memory is optional and can be viewed, edited or deleted at any time in settings.

Atlas also supports natural language commands, meaning you could type something like «reopen the shoes I looked at yesterday» or «clean up my tabs» and the browser should respond accordingly.

Read more: OpenAI Plans to Allow Erotica and Change Mental Health Restrictions for Adult Users

Agent mode in Atlas preview 

OpenAI also previewed agent mode, which lets ChatGPT take limited actions on behalf of the user — such as booking travel, ordering groceries or gathering research. The company says the mode is faster than standard ChatGPT and comes with new safeguards to keep users in control. 

Agent mode is available to Plus and Pro subscribers, and is available in beta for Business users.

«In the same way that GPT-5 and Codex are these great tools for vibe coding, we believe we can start in the long run to have an amazing tool for vibe lifing,» Will Ellsworth, the research lead for agent mode in Atlas, said during the livestream. «So delegating all kinds of tasks both in your personal and professional life to the agent in Atlas.»

How to get started with ChatGPT Atlas

To get started, you’ll first download Atlas at chatgpt.com/atlas. When you open Atlas for the first time, you’ll need to sign in to your ChatGPT account. 

From there, you can import your bookmarks, saved passwords and browsing history from your current browser. 

Continue Reading

Technologies

Amazon Will Pay $2.5 Billion for Misleading Customers Into Amazon Prime Subscriptions

Amazon settles its FTC lawsuit, and agrees to pay billions for «tricking» customers into Prime subscriptions.

In September, Amazon settled its case with the Federal Trade Commission over whether it had misled customers who signed up for Amazon Prime. The $2.5 billion settlement is one of the largest consumer protection settlements in US history, and while Amazon did not admit to wrongdoing, it’s still changing things.

The FTC said $1.5 billion will go into a fund to repay eligible subscribers, with the remaining $1 billion collected as a civil penalty. The settlement requires Amazon to add a «clear and conspicuous» option to decline Prime during checkout and to simplify the cancellation process.

«Amazon and our executives have always followed the law, and this settlement allows us to move forward and focus on innovating for customers,» Mark Blafkin, Amazon senior manager, said in a statement. «We work incredibly hard to make it clear and simple for customers to both sign up or cancel their Prime membership, and to offer substantial value for our many millions of loyal Prime members around the world.»


Don’t miss any of our unbiased tech content and lab-based reviews. Add CNET as a preferred Google source.


Why was the FTC suing Amazon?

The FTC filed suit against Amazon in 2023, accusing it of using «dark patterns» to nudge people into Prime subscriptions and then making it too hard to cancel. The FTC maintained Amazon was in violation of Section 5 of the FTC Act and the Restore Online Shoppers’ Confidence Act

«Specifically, Amazon used manipulative, coercive or deceptive user-interface designs known as ‘dark patterns’ to trick consumers into enrolling in automatically renewing Prime subscriptions,» the FTC complaint states.

Who is eligible for Amazon’s big payout?

Amazon’s legal settlement is limited to customers who enrolled in Amazon Prime between June 23, 2019, and June 23, 2025. It’s also restricted to customers who subscribed to Prime using a «challenged enrollment flow» or who enrolled in Prime through any method but were unsuccessful in canceling their memberships.

The FTC called out specific enrollment pages, including Prime Video enrollment, the Universal Prime Decision page, the Shipping Option Select page and the Single Page Checkout. To qualify for a payout, claimants must also not have used more than 10 Amazon Prime benefits in any 12-month period.

Customers who signed up via those challenged processes and did not use more than three Prime benefits within one year will be paid automatically by Amazon within 90 days. Other eligible Amazon customers will need to file a claim, and Amazon is required to send notices to those people within 30 days of making its automatic payments.

Customers who did not use a challenged sign-up process but instead were unable to cancel their memberships will also need to file claims for payment.

How much will the Amazon payments be?

Payouts to eligible Amazon claimants will be limited to a maximum of $51. That amount could be reduced depending on the number of Amazon Prime benefits you used while subscribed to the service. Those benefits include free two-day shipping, watching shows or movies on Prime Video or Whole Foods grocery discounts.

Continue Reading

Technologies

This Rumored Feature Could Make NotebookLM Essential for Work as Well as School

NotebookLM takes another step toward being the do-it-all AI tool for work and school.

Since it launched, NotebookLM has been aimed at students. While just about anyone can use the AI tool to some benefit, it’s a great study buddy thanks to an assortment of features for the classroom. But a promising new feature may help with your next work presentation: Slides.

Powered by Gemini, NotebookLM can help you brainstorm ideas and generate audio or video overviews. That sounds like most AI tools, but NotebookLM is different. You can provide it with your own material — documents, websites, YouTube videos and more — and it’ll only use those sources to answer your questions and generate content. Adding a slide generator to such a tool would be a solid, professional power-up. 


Don’t miss any of our unbiased tech content and lab-based reviews. Add CNET as a preferred Google source.


Google already has its own slide deck creation tool, but NotebookLM could make it even easier to create them. Using your uploaded sources and the recently integrated Nano Banana image generator, the ability to create a slide deck on the fly could soon be on its way. 

The tech and AI tool-focused site Testing Catalog recently spotted an unreleased and incomplete Slide tool. Not all of the features seem to be available, but it’d be easy to assume you’ll be able to create a slide deck based on your uploaded documents with just a few clicks. It’ll also likely allow you to further customize the deck by giving NotebookLM specific instructions and topics within your sources to focus on. 

That’s not all, though. Another, similar feature might also be on the way. Also spotted was an option to generate an infographic — allowing you to create a visual chart or image based on your data sources. We’ll have to wait and see when either of these features goes live, but NotebookLM remains a robust tool that has little competition, and I expect it’ll only get better. 

Continue Reading

Trending

Copyright © Verum World Media