Skip to main content
Crowdfunding
Python + AI for Geeks
Practice

What is GPT?

GPT stands for Generative Pre-trained Transformer, a type of generative AI model that is pre-trained on large datasets based on the transformer model introduced by a team at Google in 2017.

Transformer: An efficient AI architecture that can simultaneously understand the relationships between words in a sentence


Nowadays, GPT is used for almost every text-centric task, including document summarization, translation, writing, and coding.

In this lesson, we'll explore what GPT means and how it came to be so prominent.


What does GPT stand for?

GPT is an acronym for the following three words:

  • Generative: The AI model can generate text.

  • Pre-trained: It is trained in advance using a large amount of data.

  • Transformer: It utilizes the transformer architecture.

In essence, GPT represents a "pre-trained text generation model based on the transformer architecture."

Simply put, it's an AI that has learned from countless documents in advance to create new sentences naturally.


What makes GPT special?

GPT popularized the use of the transformer model, which can grasp the context of entire sentences at once.

Previous RNN-based models took a long time to generate text by processing sentences sequentially, but GPT is capable of understanding and generating sentences quickly and accurately by comparing all words simultaneously.

Additionally, it offers a fine-tuning feature, allowing you to retrain GPT for specific tasks like summarization and translation on top of a large pre-trained model.


In the next lesson, we'll delve deeper into the background and history of GPT's development.

Want to learn more?

Join CodeFriends Plus membership or enroll in a course to start your journey.