Skip to main content

What is GPT-4 and Why Does it Matter?

OpenAI has announced the release of its latest large language model, GPT-4. This model is a large multimodal model that can accept both image and text inputs and generate text outputs.
Updated Mar 2023  · 8 min read

The recent release of GPT-4 marks a significant milestone in the field of artificial intelligence, particularly in natural language processing. In this article, we offer a thorough analysis of its advanced capabilities and delve into the history and development of Generative Pre-trained Transformers (GPT), and what new capabilities GPT-4 unlocks.

What are Generative Pre-trained Transformers?

Generative Pre-trained Transformers (GPT) are a type of deep learning model used to generate human-like text. Common uses include 

  • answering questions
  • summarizing text
  • translating text to other languages
  • generating code
  • generating blog posts, stories, conversations, and other content types. 

There are endless applications for GPT models, and you can even fine-tune them on specific data to create even better results. By using transformers, you will be saving costs on computing, time, and other resources. 

Before GPT

The current AI revolution for natural language only became possible with the invention of transformer models, starting with Google's BERT in 2017. Before this, text generation was performed with other deep learning models, such as recursive neural networks (RNNs) and long short-term memory neural networks (LSTMs). These performed well for outputting single words or short phrases but could not generate realistic longer content.

BERT's transformer approach was a major breakthrough since it is not a supervised learning technique. That is, it did not require an expensive annotated dataset to train it. BERT was used by Google for interpreting natural language searches, however, it cannot generate text from a prompt.



Transformer architecture | GPT-1 Paper

In 2018, OpenAI published a paper (Improving Language Understanding by Generative Pre-Training) about using natural language understanding using their GPT-1 language model. This model was a proof-of-concept and was not released publicly.



Model performance on various tasks | GPT-2 paper

The following year, OpenAI published another paper (Language Models are Unsupervised Multitask Learners) about their latest model, GPT-2. This time, the model was made available to the machine learning community and found some adoption for text generation tasks. GPT-2 could often generate a couple of sentences before breaking down. This was state-of-the-art in 2019.



Results on three Open-Domain QA tasks | GPT-3 paper

In 2020, OpenAI published another paper (Language Models are Few-Shot Learners) about their GPT-3 model. The model had 100 times more parameters than GPT-2 and was trained on an even larger text dataset, resulting in better model performance. The model continued to be improved with various iterations known as the GPT-3.5 series, including the conversation-focused ChatGPT.

This version took the world by storm after surprising the world with its ability to generate pages of human-like text. ChatGPT became the fastest-growing web application ever, reaching 100 million users in just two months.

You can learn more about GPT-3, its uses, and how to use it in a separate article.

Learn how to use ChatGPT with our Introduction to ChatGPT course

Introduction to ChatGPT Course

Get Started with ChatGPT

Start now

What’s New in GPT-4?

GPT-4 has been developed to improve model "alignment" - the ability to follow user intentions while also making it more truthful and generating less offensive or dangerous output. 

Performance improvements

As you might expect, GPT-4 improves on GPT-3.5 models regarding the factual correctness of answers. The number of "hallucinations," where the model makes factual or reasoning errors, is lower, with GPT-4 scoring 40% higher than GPT-3.5 on OpenAI's internal factual performance benchmark.

It also improves "steerability," which is the ability to change its behavior according to user requests. For example, you can command it to write in a different style or tone or voice. Try starting prompts with "You are a garrulous data expert" or  "You are a terse data expert" and have it explain a data science concept to you. You can read more about designing great prompts for GPT models here

A further improvement is in the model's adherence to guardrails. If you ask it to do something illegal or unsavory, it is better at refusing the request.

Using Visual Inputs in GPT-4

One major change is that GPT-4 can use image inputs (research preview only; not yet available to the public) and text. Users can specify any vision or language task by entering interspersed text and images.

Examples showcased highlight GPT-4 correctly interpreting complex imagery such as charts, memes, and screenshots from academic papers. 

You can see examples of the vision input below.




GPT-4 Performance Benchmarks

OpenAI evaluated GPT-4 by simulating exams designed for humans, such as the Uniform Bar Examination and LSAT for lawyers, and the SAT for university admission. The results showed that GPT-4 achieved human-level performance on various professional and academic benchmarks.


OpenAI also evaluated GPT-4 on traditional benchmarks designed for machine learning models, where it outperformed existing large language models and most state-of-the-art models that may include benchmark-specific crafting or additional training protocols. These benchmarks included multiple-choice questions in 57 subjects, commonsense reasoning around everyday events, grade-school multiple-choice science questions, and more.

OpenAI tested GPT-4's capability in other languages by translating the MMLU benchmark, a suite of 14,000 multiple-choice problems spanning 57 subjects, into various languages using Azure Translate. In 24 out of 26 languages tested, GPT-4 outperformed the English-language performance of GPT-3.5 and other large language models.

Overall, GPT-4's more grounded results indicate significant progress in OpenAI's effort to develop AI models with increasingly advanced capabilities.

How to Gain Access to GPT-4

OpenAI is releasing GPT-4’s text input capability via ChatGPT. It is currently available to ChatGPT Plus users. There is a waitlist for the GPT-4 API. 

Public availability of the image input capability has not yet been announced. 

OpenAI has open-sourced OpenAI Evals, a framework for automated evaluation of AI model performance, to allow anyone to report shortcomings in their models and guide further improvements. 

Take it to the Next Level

In the meantime, you can read more resources on GPT-4, ChatGPT, and AI in the following resources: 


5 More Things Business Leaders Need to Know About Machine Learning

Dive deeper into what you need to know about machine learning to sustainably grow your data function and your company at large.
Hugo Bowne-Anderson's photo

Hugo Bowne-Anderson

7 min

Becoming a Kaggle Grandmaster

Jean-Francois Puget is a Distinguished Engineer at NVIDIA and a 3x Kaggle Grandmaster. We talk competitive machine learning and computing on GPUs.
Adel Nehme's photo

Adel Nehme

50 min

Stock Market Predictions with LSTM in Python

Discover Long Short-Term Memory (LSTM) networks in Python and how you can use them to make stock market predictions!
Thushan Ganegedara's photo

Thushan Ganegedara

30 min

Demystifying Mathematical Concepts for Deep Learning

Explore basic math concepts for data science and deep learning such as scalar and vector, determinant, singular value decomposition, and more.
Avinash Navlani's photo

Avinash Navlani

11 min

Deduce the Number of Layers and Neurons for ANN

There is an optimal number of hidden layers and neurons for an artificial neural network (ANN). This tutorial discusses a simple approach for determining the optimal numbers for layers and neurons for ANN's.
Ahmed Gad's photo

Ahmed Gad

9 min

Natural Language Processing Tutorial

Learn what natural language processing (NLP) is and discover its real-world application, using Google BERT to process text datasets.
DataCamp Team's photo

DataCamp Team

13 min

See MoreSee More