Day 6 Update: OpenAI Advanced Model has Vision Now & ChatGPT integrates in iOS 18.2

Day 6 Update: OpenAI Advanced Model has Vision Now & ChatGPT integrates in iOS 18.2

ChatGPT's New Voice Model with Vision Capabilities

Discover ChatGPT's Advanced Voice Model with Vision Capabilities

Welcome to an exciting update in the world of AI. OpenAI has released its most advanced version of ChatGPT, now featuring a powerful voice model that also comes with vision capabilities. In this video, I demonstrate how you can engage with ChatGPT’s voice model, show it images or your mobile screen, and receive detailed, real-time feedback.

What is ChatGPT's Voice Model with Vision?

This new model allows you to interact with ChatGPT by speaking to it while enabling it to see and analyze visual inputs. Whether you are sharing an image, uploading a photo, or showing your mobile screen, the model can process what it sees and provide detailed insights.

Live Demonstration of Vision Features

In this video, I showcased how ChatGPT's vision and voice work together by testing the following:

  • Analyzing a Thumbnail: I displayed a YouTube thumbnail from my mobile screen and asked ChatGPT for feedback on improving it. The AI provided actionable tips, like adding text overlays and enhancing clarity.
  • Engaging with My Studio Setup: By showing my studio through the camera, I asked the AI to identify equipment such as microphones and lighting. ChatGPT accurately identified details, including microphone brands.
  • Live Screen Sharing: I demonstrated the ability to share my mobile screen while asking ChatGPT to analyze visual content directly. This feature allows for seamless integration with real-time visuals.

How Does It Work?

The process of using the voice model with vision capabilities is simple and intuitive:

  1. Activate ChatGPT's voice model.
  2. Show an image, upload a photo, or share your screen during the interaction.
  3. Ask questions or seek feedback on what the AI is seeing.
  4. Receive detailed, actionable responses based on the visual input.

Why This Update is a Game-Changer

ChatGPT’s voice and vision integration unlocks endless possibilities:

  • Creators can receive instant feedback on images, thumbnails, or visual designs.
  • Professionals can analyze charts, diagrams, or documents during live conversations.
  • Users can engage hands-free, combining visual tasks with voice-driven workflows.

This update brings a new level of interactivity, bridging the gap between seeing and understanding in AI interactions.

Apple Intelligence Integration

As an added bonus, ChatGPT now integrates with Apple Intelligence on iOS 18.2. If Apple’s built-in model cannot answer complex queries, Siri seamlessly redirects the request to ChatGPT for a response. I also walk through this integration in the video and show how it works in real time.

Watch the Full Demo

To see the new vision and voice capabilities in action, watch the full video on my channel. You’ll see live examples of how these features work, including direct interactions with ChatGPT to analyze images and provide feedback.

Subscribe to my channel for more AI updates and practical demonstrations:

Click here to subscribe

For exclusive tips, insights, and updates, join my newsletter here:

Sign up for the newsletter

Stay tuned as I explore more innovations in AI and technology. Exciting times ahead!

Back to blog