GPT-4o (“o” for “omni”) is a step towards much more natural human-computer interaction—it accepts as input any combination of text, audio, and image and generates any combination of text, audio, and image outputs. It can respond to audio inputs in as little as 232 milliseconds, with an average of 320 milliseconds, which is similar to human response time(opens in a new window) in a conversation. It matches GPT-4 Turbo performance on text in English and code, with significant improvement on text in non-English languages, while also being much faster and 50% cheaper in the API. GPT-4o is especially better at vision and audio understanding compared to existing models.

Prior to GPT-4o, you could use Voice Mode to talk to ChatGPT with latencies of 2.8 seconds (GPT-3.5) and 5.4 seconds (GPT-4) on average. To achieve this, Voice Mode is a pipeline of three separate models: one simple model transcribes audio to text, GPT-3.5 or GPT-4 takes in text and outputs text, and a third simple model converts that text back to audio. This process means that the main source of intelligence, GPT-4, loses a lot of information—it can’t directly observe tone, multiple speakers, or background noises, and it can’t output laughter, singing, or express emotion.

GPT-4o’s text and image capabilities are starting to roll out today in ChatGPT. We are making GPT-4o available in the free tier, and to Plus users with up to 5x higher message limits. We’ll roll out a new version of Voice Mode with GPT-4o in alpha within ChatGPT Plus in the coming weeks.

  • Sabata11792
    link
    fedilink
    1
    edit-2
    6 months ago

    I been playing with the Mistral 7b models. About the most my hardware can reasonably run… so far. Would love to add vision and voice but I’m just happy it can run.

    • @ProfessorProteus
      link
      English
      16 months ago

      I’ve been wanting to run that one on my hardware but GPT4All just refuses to start its GUI. The only thing is a “chat.exe” that sits idle in the task manager. And this is an issue I’ve seen reported in their Github from several users, on both Win 10 and 11.

      Have you find success with that frontend, or are you using one that actually works? I haven’t researched any others since this issue has me a little burnt out.

      • Sabata11792
        link
        fedilink
        16 months ago

        I use TextgenWebui and sometimes Kobold. I can only run it with 4bit quant enabled since I’m just short on VRam to fully load the model.

        Text gen runs a server you access though the web browser instead of a desktop app.

        I haven’t tried GPT4all.

        • @ProfessorProteus
          link
          English
          16 months ago

          I’ll definitely give the web UI a shot since I’m already quite familiar with A1111 and it seems they’re trying to recreate the same look and feel. Thank you for the info!

          • Sabata11792
            link
            fedilink
            16 months ago

            No problem. It’s fairly easy to figure out in a few minutes if you know Auto1111. Getting your model to actually load may need some tweaking, but I managed to just trial and error it.