Google has announced that its AI, Gemini, will soon be capable of handling complex, multi-step tasks directly on your smartphone, commencing with actions like ordering food or hailing a ride.
This advanced functionality, initially rolling out to devices such as the Pixel 10, Pixel 10 Pro, and the recently unveiled Samsung Galaxy S26 phones, echoes features Apple first showcased for Siri at its 2024 Worldwide Developers Conference. However, Apple subsequently delayed those planned features in March 2025, and they currently remain unreleased.
During an onstage presentation, Sameer Samat, Google’s president of Android, demonstrated how Gemini's new agentic capabilities could streamline a pizza dinner order originating from a busy family group chat. Samat instructed Gemini to analyze the chat thread to discern preferences, then proceed to place the order via a delivery application. A prerecorded video displayed Gemini interpreting individual requests from the group chat and presenting them in a dedicated window. The user then, through a voice command, directed Gemini to finalize the order with a specific pizzeria. Gemini subsequently navigated through GrubHub, preparing the order directly onscreen. Once the order was assembled, Gemini issued an alert, allowing the user to review and ultimately confirm the submission.
While the demonstrated scenario might seem straightforward enough to execute manually within a delivery app or even by directly contacting the pizzeria, this development represents a potentially significant moment for agentic AI. Following Gemini's recent integration of auto-browsing capabilities within Chrome, extending similar functionality directly into Android feels like a logical progression. Google is clearly aiming to position Gemini as an invaluable helpful agent and productivity partner, rather than merely a chatbot or a collection of AI models.
Assuming Google delivers on its promise of a "soon" launch for these agentic Gemini features, and absent an unexpected counter-move from Apple, Google appears set to precede Apple in deploying functionalities akin to its impressive Apple Intelligence demos from WWDC 2024—which were also presented via prerecorded videos. One feature Apple showcased involved Siri's ability to comprehend on-screen content and execute actions, such as extracting an address from a Messages thread to update a contact card. Apple also illustrated Siri's capacity to perform tasks both within and across applications, and even understand personal context, exemplified by retrieving flight information from an email in response to a query about a parent's arrival time.
Nearly two years later, none of those Apple Intelligence features are yet available. When Apple announced the delays, the company even withdrew an advertisement that had showcased these capabilities. Furthermore, based on reporting from Bloomberg, some of these features may not materialize until iOS 27.
Naturally, numerous questions persist regarding Gemini’s new capabilities. Their actual deployment is crucial, and real-world testing will be necessary to ascertain if they are as useful and functional as advertised—especially since Google is labeling this initial launch a “beta,” suggesting potential refinements may be needed. A key unknown is the extent to which developers will permit Gemini to browse through their applications on behalf of users, a challenge Verge editor-in-chief Nilay Patel refers to as "the DoorDash problem." Google has stated that Gemini will be able to work in “select rideshare and food apps.”
Nevertheless, Google appears to have significantly advanced ahead of Apple in this domain, placing increased pressure on Apple to accelerate its own development efforts.
The Editorial Staff at AIChief is a team of professional content writers with extensive experience in AI and marketing. Founded in 2025, AIChief has quickly grown into the largest free AI resource hub in the industry.