Apple doesn'tcurrently allow iPhone users to change the Side button's Siri functionality to another assistant, but owners of iPhone 15 Pro and newer models can assign ChatGPT to the Action button instead. Keep reading to learn how.

OpenAI's free ChatGPT app for iPhone lets you interact with the chatbot using text or voice. What's more, if you assign voice mode directly to the Action button, you can use it to jump straight into a spoken conversation, giving you quick, handsfree access to a far more capable assistant.
- Install the ChatGPT app, then sign into your account or create a new one.
- Next, open the Settings app.
- TapAction Button.
- Swipe toControls, then tap the two chevrons beside the currently assigned control.
- Using the control search field, type "ChatGPT."
- Select the controlOpen ChatGPT Voice.

A long press of the Action button will now open ChatGPT's voice mode. The first time you activate it, the app may request microphone access. TapAllow to proceed. After that, you can begin speaking immediately.
Arecent update means voice conversations now take place inside the same chat window as text-based prompts, instead of switching to a separate voice-only interface. Responses appear in real time, combining spoken output with on-screen text and any visuals the model generates. This keeps your conversation's context intact and makes switching between typing and speaking smoother.
Continue the Conversation Anywhere
You can also leave the ChatGPT app during a voice session without ending it. When you swipe away, the conversation continues and appears in the Dynamic Island as long as the assistant is listening and preparing a response. To stop, tap the Dynamic Island to return to the app, then tap End.
Note that while this setup gives you a fast, physical shortcut to a richer, more context-aware assistant, ChatGPT can't perform Siri-like system actions like accessing your calendar or setting an alarm.
Top Rated Comments
Siri isn't going away.
It won’t. Siri though, might. Well not in name perhaps, but it’s very likely it will just be a skin for Google Gemini.
The report on the street is that Apple is going to license a separate installation of Gemini that will reside on Apple's PPC (Private Cloud Compute) servers to power Siri. Google would have zero access to this installation, so no data scraping will occur.
Apple has been wise to not rush into the LLM-powered AI Assistant era until the technology has matured and the compute costs come down. That's now happening.
What Apple should've done is gone full bore into the SLM (Small Language Model) architecture instead of "one model to rule them all". Have specialized models for different tasks, with Siri being a "conductor", but this technology also didn't exist in a usable form until recently.
While the frontier labs are releasing new models are break-neck speeds, we're only just now — this week, in fact — starting to see models that would be able to meet the needs of Siri.
Great time to be alive.
Popular Stories

Here's Why the Apple Store is Going Down

iPhone Pocket is Now Completely Sold Out Worldwide

Best Black Friday Streaming Deals - Save Big on Apple TV, Disney+, Hulu, and More

Apple and Intel Rumored to Partner on Mac Chips Again in a New Way
Foldable iPhone to Debut These Three Breakthrough Features

iPhone Air Flop Sparks Industry Retreat From Ultra-Thin Phones

The Best Black Friday iPhone Deals Still Available














