DALL-E 3 and Vision in ChatGPT are Separate Features
DALL-E 3 image of a "photo realistic Kingfisher in nature"

DALL-E 3 and Vision in ChatGPT are Separate Features

I was confused how ChatGPT's DALL-E 3 image generating feature and its new vision feature worked together.

Instinctively, you group those two capabilities together in your mind and expect to use them interchangeably.

I also saw other savvy tech people asking about how this works.

The answer is, they're separate features.

To generate an image with DALL-E 3 in your ChatGPT Plus account:

  • Start a new chat
  • Select the GPT-4 model
  • In the dropdown menu under the GPT-4 model, select DALL-E 3
  • In your chat window, ask GPT-4 to generate any type or style of image
  • Mobile and desktop work the same

To upload an image to be seen and interpreted by GPT-4 (requires ChatGPT Plus):

  • Start a new chat
  • Select the GPT-4 model
  • In the dropdown menu under the GPT-4 model, select "Default"
  • In your chat window, click the image icon to the left of "Send a message"
  • Upload your image and ask anything you want about it
  • This works essentially the same way in the mobile app as it does in the desktop browser

How do you turn these two features on?

  • You don't turn the vision feature on; the image icon simply appears in your chat window when you have "Default" selected in the GPT-4 model dropdown
  • You do turn DALL-E 3 on in your account's "Settings & Beta", and then you'll see the "DALL-E 3" option in the GPT-4 model dropdown menu

These two features are flat out incredible. Enjoy!

Article content
Use the Default model for ChatGPT's vision feature.
Article content
Upload an image to the Default model to discuss it with ChatGPT.



Bradley Jefferson

Full stack Immersive Product Manager (web, mobile, XR, AI/ML) | XR Independent Producer | GenAI Interactive Filmmaker

1y

I had the biggest challenge with spelling and Dall-e. Very frustrating. Even when you spell the words for it. Did you have the same experience? My use case was creating illustrative movie posters.

Separation of features provides more versatility and functionality for different AI use cases and applications 🚀 curious how often you use image recognition Brooks ?

Mike Fraietta

Co-founder at Gargoyle Systems, Partnerships at Nokbox.

1y

Okay I was trying the last few days and wondering how to do it. Needed this! Will report back.

Brooks Jordan Very Informative. Thank you for sharing.

To view or add a comment, sign in

More articles by Brooks Jordan

Insights from the community

Others also viewed

Explore topics