When I logged into OpenAI’s website on Monday morning to continue testing the new version of ChatGPT powered by GPT-4, the chatbot didn’t try to sabotage my relationship, write my emails, or unleash my creativity—it simply didn’t work. Demand is high, and the company is experiencing occasional outages. Greg Brockman, an OpenAI cofounder and president, was upfront about the model’s imperfection in a recent livestream. He also reminded listeners that they were not without blemishes themselves.
Generative AI is the focal point for many Silicon Valley investors after OpenAI’s transformational release of ChatGPT late last year. The chatbot uses extensive data scraped from the internet and elsewhere to produce predictive responses to human prompts. It was previously powered by the GPT-3.5 language model. While that version remains online, an algorithm called GPT-4 is now available with a $20 monthly subscription to ChatGPT Plus.
If you’re considering that subscription, here’s what you should know before signing up, with examples of how outputs from the two chatbots differ.
The core service you pay for with ChatGPT Plus is access to GPT-4. Even after paying $20 a month, you aren’t guaranteed a specific number of prompts from the GPT-4 model per day. OpenAI says clearly that the company will change the maximum number of allowed prompts at any time. While I was testing it out on a Friday afternoon, the cap was set at 50 messages for four hours. When I returned on Monday morning, the site was glitchy and the cap was lowered to 25 messages for three hours.
OpenAI claims that ChatGPT Plus subscribers can avoid getting bumped out of the chatbot during peak usage hours and receive faster responses. But users should know that they might lose access to ChatGPT during some outages. In addition, the current iteration of the GPT-4 model often takes longer to respond to prompts than GPT-3.5 does.
You can experiment with a version of GPT-4 for free by signing up for Microsoft’s Bing and using the chat mode.
Many aspects of GPT-4 remain opaque. OpenAI has not shared many details about GPT-4 with the public, like the model’s size or specifics about its training data. Subscribing to ChatGPT Plus does not yet grant access to the image-analysis capabilities recently previewed by the company.
With that in mind, are answers from the GPT-3.5 and GPT-4 models noticeably different? OpenAI’s website reads, “GPT-4 excels at tasks that require advanced reasoning, complex instruction understanding, and more creativity.” It seems like the new model performs well in standardized situations, but what if we put it to the test? Below are the two chatbots’ initial, unedited responses to three prompts we crafted specifically for that purpose.
GPT-3.5:
Knock knock.
Who's there?
Boo.
Boo who?
Don’t cry, it’s just a joke! (cue Pulitzer Prize-worthy laughter)