How do language-based AIs, such as GPT, work?
Table of contents:
What does GPT stand for?
How does GPT work?
How can GPT feel so smart?
Why is GPT often referred to as a “model?”
Is GPT smart enough to be sentient?
What’s the difference between GPT and ChatGPT?
Who owns GPT?
Can I download ChatGPT and try it for free?
Could GPT dominate humanity at some point?
Learn how to build a simple language model
This article aims at demystifying GPT in the simplest way possible and give you a high-level understanding of large language models (or LLMs).
That way, you will be able to use them more effectively in your next project or when “chatting” with them (through ChatGPT for instance).
GPT stands for Generative Pre-trained Transformers.
This article will tell you everything you need to know about why this Artificial Intelligence is generative, why it’s been pre-trained, and what is a Transformer, in the easiest way I possibly can.
Here’s a simplified way to see it: At the root of GPT is a word prediction algorithm (named “Transformer”) that is based on the patterns it notices from its training data.
All it sees are numbers. They’re also called “tokens.” They’re the numerical representation of the words that fed its knowledge base.
GPT has been trained with such an astronomical amount of data that, at some point, it started to show signs of (artificial) intelligence.
Imagine something that’s been forced to digest a huge amount of knowledge and can’t help but become smarter thanks to it.
However, don’t try to think of it as a human being or even a particularly smart animal. You’ll have a hard time making sense of this.
GPT won’t learn anything else beyond its training data. It doesn’t work like our brains. I don’t recommend you to see it as anything else than a next word prediction algorithm.
GPT is called a model because it has been trained to model the statistical patterns of language.
More specifically, it has learned to predict the probability of a word given the previous one in a sentence, which allows it to generate text that is similar to what humans do.
For now, GPT is just a language model that gives the illusion of intelligence and is not sentient. It isn’t aware of itself and doesn’t have desires. It’s just a next word prediction algorithm and nowhere near a conscious being.
ChatGPT is just a web application that wraps the GPT model into a user-friendly interface so anybody can interact with it.
GPT is owned by OpenAI.
When I discovered the company, I was surprised to learn that OpenAI began as a non-profit.
It was founded in December 2015, with the goal of promoting and developing friendly AI in a way that benefits humanity as a whole.
It has since transitioned to a capped-profit model and has made significant strides in AI research and development, notably creating advanced AI language models like GPT-3 and GPT-4, and engaging in various AI-related partnerships and applications. (The most notable being Microsoft that integrated GPT into Bing.)
You can also use ChatGPT on any platform via its website.
AI could change humanity for the better.
But some people could use it for harmful matters.
Experts even say it’s possible that AI could try to control humanity at some point, even if we can’t imagine its motives yet.
Currently, though, GPT is light years away from being able to do that.
Andrej Karpathy, one of OpenAI’s co-founders, created an incredible video showing how to build a simplified version of GPT.
This tiny GPT predicts the next character instead of the next word and is based on a small dataset. That makes everything less computationally intensive, and you can focus on learning.
Are you a developer? If so, are you ready to leverage GPT in your applications?
Then, check out my articles on this matter: