Sat. May 18th, 2024

Key Takeaways

  • GPT-4o is half the cost and twice the speed of GPT-4 Turbo, with added multimodal capabilities.
  • GPT-4o can interpret nonverbal elements, respond with an emotional range, and handle large amounts of data.
  • GPT-4o is available to free and paid users, with a higher usage limit for paid users; future features will be released gradually.

The march of progress for generative AI is turning into more of a sprint, and OpenAI’s announcement and release of GPT-4o “Omni” promises to leap ChatGPT and its related services forward in several areas. If you don’t have access to GPT-4o already, you will soon. So let’s look at what makes this new AI model different from what’s come before.

GPT-4o vs. GPT-4 Turbo and GPT-3.5

We’ve covered the difference between GPT3.5 and GPT-4 Turbo in detail already, but the short version is that GPT-4 is significantly smarter than GPT-3.5. It can understand much more nuance, produce more accurate results, and is significantly less prone to AI hallucinations. GPT-3.5 is still very much relevant since it’s fast, available for free, and can still handle most everyday tasks with ease. As long as you’re aware that it’s much more likely to give you incorrect information.

GPT-4 Turbo had been the flagship model until the arrival of GPT-4o. This model was only available for ChatGPT Plus subscribers and came with all the fun toys OpenAI developed, such as custom GPTs and live web access.

Before we get into the actual features and capabilities, an important fact is that (according to OpenAI) GPT-4o is half the cost and twice the speed of GPT-4. This is likely why GPT-4o is available to both free and paid users, although paying users will get five times the usage limit for GPT-4o.

It’s broadly just as smart as GPT-4 Turbo, so overall this advancement is more about efficiency. However, there are some obvious and not so obvious advancements that go beyond simply being faster and cheaper.

What Can GPT-4o Do?

The big buzzword with GPT-4o is “multimodal”, which is to say that GPT-4o can operate on audio, images, video, and text. Well, so could GPT-4 Turbo, but this time it’s different under the hood.

OpenAI says that it’s trained a single neural net on all modalities at once. With the older GPT-4 Turbo, when using voice mode, it would use one model to convert your spoken words to text. Then GPT-4 would interpret and respond to that text, which was then converted back into a synthesized AI voice.

With GPT-4o it’s all one model, which has numerous knock-on effects on performance and abilities. OpenAI claims that the response time when conversing with GPT-4o is now only a few hundred milliseconds, about the same as a real-time conversation with another human. Compare this to the 3-5 seconds it took for older models to respond, and it’s a major leap.

Apart from being much snappier, it also means that GPT-4o can now also interpret nonverbal elements of speech, such as the tone of your voice, and its own response now has an emotional range. It can even sing! In other words, OpenAI has given GPT-4o some measure of affective computing ability.

The same sort of efficiency and unification applies to text and images, as well as video. In one demo GPT-4o is shown having a real-time conversation with a human being with live video and audio. Just like having a video chat with a human, it appears that GPT-4o can interpret what it sees through the camera, and make some pretty incisive inferences. ChatGPT-4o can also keep a substantially larger number of tokens in mind compared to previous models, which means it can apply its intelligence to much longer conversations and larger amounts of data. This will likely make it more useful for tasks like helping you write a novel, as one example.

Now, as of this writing, not all of these features are available to the public yet, but OpenAI has stated that it will roll them out in the weeks following the initial announcement and release of the core model.

How Much Does GPT-4o Cost?

GPT-4o is available to both free and paid users, although paying users will get five times the usage limit for GPT-4o. As of this writing, the cost of ChatGPT Plus is still $20 a month, and if you’re a developer well, you’ll have to look up the API costs for your needs, but GPT-4o should be much cheaper compared to other models.

How to Use GPT-4o

As I mentioned above, GPT-4o is available to both free and paid users, but not all of its features are online right away. So, depending on when you read this, the exact things you can do with it will vary. That said, using GPT-4o is dead simple. If you’re using the ChatGPT website as a ChatGPT Plus subscriber, just click the drop down menu that displays the current model name and select GPT-4o.

The ChatGPT-4o selection menu.

If you are a free user, you will be defaulted to ChatGPT-4o until you run out of your allocation. At which point, you’ll be dropped back to version 3.5 until you have more requests available.

As of this writing, ChatGPT-4o is not yet available in the free tier, so the above is based on OpenAI’s documentation. Things might work slightly differently when the feature rolls out to free users.

If you’re using the mobile app, simply tap the three dots and choose the model from there.

ChatGPT App GPT-4o selection menu.

Then use ChatGPT as you usually would!

When ChatGPT-4o is released in its final form, we might finally get the conversant, natural language assistant that science fiction has promised. Whether that’s a good or a bad thing I leave completely up to you, but it sure is exciting!

Source link

By John P.

Leave a Reply

Your email address will not be published. Required fields are marked *