Journalism of Courage
Advertisement
Premium

What is GPT-4 and how is it different from ChatGPT?

OpenAI has said that GPT-4 is supposedly bigger, faster, and more accurate than ChatGPT. It can now analyse image-related queries as well. Here's what to expect and how you can try it right now.

GPT-4 logoThe most noticeable change to GPT-4 is that it’s multimodal, allowing it to understand more than one modality of information. It can be fed images and asked to output information accordingly. (Image via OpenAI)
Listen to this article Your browser does not support the audio element.

AI powerhouse OpenAI announced GPT-4, the next big update to the technology that powers ChatGPT and Microsoft Bing, the search engine using the tech, on Tuesday. GPT-4 is supposedly bigger, faster, and more accurate than ChatGPT, so much so, that it even clears several top examinations with flying colours, like the Uniform Bar Exam for those wanting to practice as lawyers in the US.

The company has made the language model’s powers known in its announcement blog for the same, saying that it is more creative and collaborative than ever before. Where GPT-3.5-powered ChatGPT only accepted text inputs, GPT-4 can also use images to generate captions and analyses. But that’s only the tip of the iceberg. We explain what the new language model is, what it isn’t, and how it outsmarts its predecessor.

What is GPT-4?

GPT-4 is a large multimodal model created by OpenAI and announced on March 14, 2023. Multimodal models can encompass more than just text – GPT-4 also accepts images as input. Meanwhile, GPT-3 and GPT-3.5 only operated in one modality, text, meaning users could only ask questions by typing them out.

Aside from the fresh ability to process images, OpenAI says that GPT-4 also “exhibits human-level performance on various professional and academic benchmarks.” The language model can pass a simulated bar exam with a score around the top 10 per cent of test takers and can solve difficult problems with greater accuracy thanks to its broader general knowledge and problem-solving abilities.

For example, it can “answer tax-related questions, schedule a meeting among three busy people, or learn a user’s creative writing style.”

GPT-4 is also capable of handling over 25,000 words of text, opening up a greater number of use cases that now also include long-form content creation, document search and analysis, and extended conversations.

How is GPT-4 different from GPT-3?

Here are some of the major differences:

Story continues below this ad

*GPT-4 can ‘see’ images now: The most noticeable change to GPT-4 is that it’s multimodal, allowing it to understand more than one modality of information. GPT-3 and ChatGPT’s GPT-3.5 were limited to textual input and output, meaning they could only read and write. However, GPT-4 can be fed images and asked to output information accordingly.

If this reminds you of Google Lens, then that’s understandable. But Lens only searches for information related to an image. GPT-4 is a lot more advanced in that it understands an image and analyses it. An example provided by OpenAI showed the language model explaining the joke in an image of an absurdly large iPhone connector. The only catch is that image inputs are still a research preview and are not publicly available.

*GPT-4 is harder to trick: One of the biggest drawbacks of generative models like ChatGPT and Bing is their propensity to occasionally go off the rails, generating prompts that raise eyebrows, or worse, downright alarm people. They can also get facts mixed up and produce misinformation.

OpenAI says that it spent 6 months training GPT-4 using lessons from its “adversarial testing program” as well as ChatGPT, resulting in the company’s “best-ever results on factuality, steerability, and refusing to go outside of guardrails.”

Story continues below this ad

*GPT-4 can process a lot more information at a time: Large Language Models (LLMs) may have been trained on billions of parameters, which means countless amounts of data, but there are limits to how much information they can process in a conversation. ChatGPT’s GPT-3.5 model could handle 4,096 tokens or around 8,000 words but GPT-4 pumps those numbers up to 32,768 tokens or around 64,000 words.

This increase means that where ChatGPT could process 8,000 words at a time before it started to lose track of things, GPT-4 can maintain its integrity over way lengthier conversations. It can also process lengthy documents and generate long-form content – something that was a lot more limited on GPT-3.5.

*GPT-4 has an improved accuracy: OpenAI admits that GPT-4 has similar limitations as previous versions – it’s still not fully reliable and makes reasoning errors. However, “GPT-4 significantly reduces hallucinations relative to previous models” and scores 40 per cent higher than GPT-3.5 on factuality evaluations. It will be a lot harder to trick GPT-4 into producing undesirable outputs such as hate speech and misinformation.

*GPT-4 is better at understanding languages that are not English: Machine learning data is mostly in English, as is most of the information on the internet today, so training LLMs in other languages can be challenging.

Story continues below this ad

But GPT-4 is more multilingual and OpenAI has demonstrated that it outperforms GPT-3.5 and other LLMs by accurately answering thousands of multiple-choice across 26 languages. It obviously handles English best with an 85.5 per cent accuracy, but Indian languages like Telugu aren’t too far behind either, at 71.4 per cent. What this means is that users will be able to use chatbots based on GPT-4 to produce outputs with greater clarity and higher accuracy in their native languages.

Can you try GPT-4 right now?

GPT-4 has already been integrated into products like Duolingo, Stripe, and Khan Academy for varying purposes. While it’s yet to be made available for all for free, a $20 per month ChatGPT Plus subscription can fetch you immediate access. The free tier of ChatGPT, meanwhile, continues to be based on GPT-3.5.

However, if you don’t wish to pay, then there’s an ‘unofficial’ way to begin using GPT-4 immediately. Microsoft has confirmed that the new Bing search experience now runs on GPT-4 and you can access it from bing.com/chat right now.

Meanwhile, developers will gain access to GPT-4 through its API. A waitlist has been announced for API access, and it will begin accepting users later this month.

Zohaib is a tech enthusiast and a journalist who covers the latest trends and innovations at The Indian Express's Tech Desk. A graduate in Computer Applications, he firmly believes that technology exists to serve us and not the other way around. He is fascinated by artificial intelligence and all kinds of gizmos, and enjoys writing about how they impact our lives and society. After a day's work, he winds down by putting on the latest sci-fi flick. • Experience: 3 years • Education: Bachelor in Computer Applications • Previous experience: Android Police, Gizmochina • Social: Instagram, Twitter, LinkedIn ... Read More

Tags:
  • artificial intelligence ChatGPT Explained Sci-Tech Express Explained
Edition
Install the Express App for
a better experience
Featured
Trending Topics
News
Multimedia
Follow Us
Express PremiumFrom kings and landlords to communities and corporates: The changing face of Durga Puja
X