What is GPT-3, and how can we benefit from it?

Umut Gokbayrak
Umut Gokbayrak 07 August 2020

GPT-3 is beginning to trend and take center stage in the internet world. What makes it even more interesting is the fact that it can do a wide range of things and tasks using the same algorithm. It can code, chat, write applications, create content, write an email or blog post, play chess with you, and can even chit chat with you.

While GPT-3 comes with a lot of exciting things, we need to strike a balance between the risks posed by it and the benefits offered by GPT-3.

In this article, we take you through the following sections:

  • What is GPT-3?
  • Why it is important and special
  • What makes it different from GPT-2
  • How it changes the Machine Language, AI, and NLP world
  • How can we start using it
  • Use cases and examples

What is GPT-3?

Let us take a closer look at what GPT-3 is. GPT-3 is the third generation of OpenAI's machine learning system that uses an algorithm based on 45TB (45 terabytes) of text data. It applies machine learning to generate various types of contents, including stories, code, legal documents, and even translations based on just a few input words. This stands as the world's most impressive AI, and this is for certain reasons.

Why is it important and special?

One special thing that makes GPT-3 so important is that, so far, so good, it is the largest trained model. It has a learning parameter of 175 Billion parameters, which makes it 10 times larger than any language model ever created. No wonder GPT-3 is remarkably smart. It has the edge over other models in that it can perform tasks without lots of tuning; it only requires little textual-interactional demonstration, and the model does the rest. It is so important in accomplishing the following and even more:

  • Writing news articles - given only a title
  • Performance of up to 5-digits arithmetic with great accuracy
  • Translation of common languages (improvement compared to the GPT-2)
  • Story writing with good endings
  • Prediction of last words of sentences by contextual recognition
  • Answering of questions. Including trivial puzzles with correctness.

We can see all the uniqueness of GPT-3, but what makes it different from other NLPs and most especially what makes it different from GPT-2?

What Makes GPT-3 different from GPT-2?

OpenAI published their unsupervised Language Model, GPT-2, in February of 2019. This model was trained using the 40GB of text and enabled it to predict words in proximity. GPT-2 produces artificial text based on the model from arbitrary input. It learns from the style and condition of the text. Finally, it is built using 1.5 billion parameters.

GPT-3, on the other hand, has an outstanding 175 billion parameters and trained on 45TB of text. Although it is tailored from the GPT-2 model, it can do way more things than the GPT-2 model. Moreover, it involves reversible tokenization, pre-normalization, and an adjustable initialization.

Furthermore, GPT-3 was trained using a high-bandwidth cluster by Microsoft and utilizing it on the V100 GPU. GPT-3 is performed under three shots settings;

  1. Zero-shot
  2. One-shot
  3. Few-shot
the three settings we explore for in-context learning

How does GPT-3 change the ML, AI, and NLP world?

There has not been anything this big as far AS machine language, and Artificial intelligence is concerned. We have seen GPT-2 with 1.5 billion parameters, we also encountered NVIDIA's Megatron with 8 billion parameters, and this was soon eclipsed by Microsoft's Turing NLG with 17 billion parameters. Now OpenAI has changed the playing field with GPT-3 utilizing a whooping 175 billion parameters. Current NLP models, to a large extent, still have to learn using a few examples. This has been changed by the entry of GPT-3, which requires very little demonstration.

gtp-3 vs other models in terms of accuracy

The image above shows how GPT-3 has performed compared to other models. It is a game-changer in the NLP world

How can we start using it?

In contrast to the previous release by OpenAi; GPT-2, which was an open-source model, GPT-3 is only available as an API release. To get started, you would need to apply to be whitelisted by the company. They state three reasons for releasing an API instead of an open-source model.

  1. To help pay for the cost of ongoing AI research, policy, and safety.
  2. Due to the large nature of the model, it can be very expensive to run and deploy. This makes it very difficult for smaller organizations and businesses to access at this stage. (but they are hopeful that the API would become more accessible by all)
  3. To curb misuse of the technology since it is difficult to predict the entirety of uses for the model. It is, therefore, safer to release via API and then increase access over time.

Use cases and Examples

There are a lot of use cases for which GPT-3 is applicable. It ranges from sending bulk company email to writing blog posts to writing codes. The list is endless. Here is a list of some uses of GPT-3.

  • Creation of apps and layout tools
  • Search and data analysis
  • Program generation and program analysis
  • Text generation
  • Content creation
  • General reasoning and mathematics
  • Chess games IVR designs using natural language
  • Patient diagnosis from clinical vignettes
  • Turning legal text into plain English
Let’s take a look at some of these examples

Creation of apps and layout tools

Jordan Singer tested GPT-3 to create an app by only providing a description for GPT-3. You can find it here. Also, Sharif Shameem built a functional react app by only describing what he wanted to GPT-3.

Turning Legal text into plain English

Another user, Michael, trained GPT-3 to turn legal text into simple and plain English without using a code. You can check out this example here.

Content creation

Content creation with GPT-3 is seamless and has received great applaud from lots of testers. For instance, Merzmensch twitted about his first try with the app. He sought for a Shakespearean poem and got something splendid. Take a look here.

Some other great examples:

In conclusion,

OpenAI's GPT-3 is a huge step towards better application for NLP in the AI space. No doubt it is going to help a lot of industries. All it demands is the right applications and right combinations with the model, and you are good to go. GPT-3 might just be the next big thing for the world with its groundbreaking innovation and enormous benefits.

Umut Gokbayrak
Written by

Umut Gokbayrak

I turn ideas into software products