Jamie Dimon warned of economic threats as his bank posted its biggest annual profit ever
Subtitles
  • Off
  • English

AI has a lot of terms. We've got a glossary for what you need to know

AI has a lot of terms. We've got a glossary for what you need to know

GPU? TPU? LLM? All the important AI vocabulary to know

We may earn a commission from links on this page.
Start Slideshow
Start Slideshow
The Open AI logo is displayed at the Mobile World Congress in Barcelona, Spain.
The Open AI logo is displayed at the Mobile World Congress in Barcelona, Spain.
Image: Joan Cros/NurPhoto (Getty Images)

When people unfamiliar with AI envision artificial intelligence, they may imagine Will Smith’s blockbuster I, Robot, the sci-fi thriller Ex Machina, or the Disney movie Smart House — nightmarish scenarios where intelligent robots take over to the doom of their human counterparts.

Advertisement

Today’s generative AI technologies aren’t quite all-powerful yet. Sure, they may be capable of sowing disinformation to disrupt elections or sharing trade secrets. But the tech is still in its early stages, and chatbots are still making big mistakes.

Still, the newness of the technology is also bringing new terms into play. What makes a semiconductor, anyway? How is generative AI different from all the other kinds of artificial intelligence? And should you really know the nuances between a GPU, a CPU, and a TPU?

If you’re looking to keep up with the new jargon the sector is slinging around, Quartz has your guide to its core terms.

Advertisement
Previous Slide
Next Slide

2 / 16

What is Generative AI?

What is Generative AI?

Jensen Huang standing in front of a display that says Generative AI and a drawing of a diagram
Nvidia CEO Jensen Huang.
Photo: Justin Sullivan (Getty Images)

Let’s start with the basics for a refresher. Generative artificial intelligence is a category of AI that uses data to create original content. In contrast, classic AI could only offer predictions based on data inputs, not brand new and unique answers using machine learning. But generative AI uses “deep learning,” a form of machine learning that uses artificial neural networks (software programs) resembling the human brain, so computers can perform human-like analysis.

Advertisement

Generative AI isn’t grabbing answers out of thin air, though. It’s generating answers based on data it’s trained on, which can include text, video, audio, and lines of code. Imagine, say, waking up from a coma, blindfolded, and all you can remember is 10 Wikipedia articles. All of your conversations with another person about what you know are based on those 10 Wikipedia articles. It’s kind of like that — except generative AI uses millions of such articles and a whole lot more.

Advertisement
Previous Slide
Next Slide

3 / 16

What is a chatbot?

What is a chatbot?

Image for article titled AI has a lot of terms. We've got a glossary for what you need to know
Image: CFOTO/Future (Getty Images)

AI chatbots are computer programs that generate human-like conversations with users, giving unique, original answers to their queries. Chatbots were popularized by OpenAI’s ChatGPT, and since then, a bunch more have debuted: Google Gemini, Microsoft CoPilot, and Salesforce’s Einstein lead the pack, among others.

Advertisement

Chatbots don’t just generate text responses — they can also build websites, create data visualizations, help with coding, make images, and analyze documents. To be sure, AI chatbots aren’t foolproof yet — they’ve made a lot of mistakes already. But as AI technology rapidly advances, so will the quality of these chatbots.

Advertisement
Previous Slide
Next Slide

4 / 16

What is a Large Language Model (LLM)?

What is a Large Language Model (LLM)?

Google Gemini.
Google Gemini.
Photo: NurPhoto/Contributor (Getty Images)

Large language models (LLMs) are a type of generative artificial intelligence. They are trained on large amounts of data and text, including from news articles and e-books, to understand and generate content, including natural language text. Basically, they are trained on a ton of text so they can predict what word comes next. Take this explanation from Google:

“If you started to type the phrase, “Mary kicked a…,” a language model trained on enough data could predict, “Mary kicked a ball.” Without enough training, it may only come up with a “round object” or only its color “yellow.” — Google’s explainer

Advertisement

Popular chatbots like OpenAI’s ChatGPT and Google’s Gemini, which have capabilities such as summarizing and translating text, are examples of LLMs.

Advertisement
Previous Slide
Next Slide

5 / 16

What is a semiconductor?

What is a semiconductor?

close up of gloved hand holding microchip
A semiconductor is also called a microchip.
Photo: Wong Yu Liang (Getty Images)

No, it’s not an 18-wheeler driver. Semiconductors, also known as AI chips, are used in electrical circuits of devices such as phones and computers. Electronic devices wouldn’t exist without semiconductors, which are made from pure elements like silicon or compounds like gallium arsenide, to conduct electricity. The name “semi” comes from the fact that the material can conduct more electricity than an insulator, but less electricity than a pure conductor like copper.

Advertisement

The world’s largest semiconductor foundry, Taiwan Semiconductor Manufacturing Company (TSMC), makes an estimated 90% of advanced chips in the world, and counts top chip designers Nvidia and Advanced Micro Devices (AMD) as customers.

Even though semiconductors were invented in the U.S., it now produces about 10% of the world’s chips, not including advanced ones needed for larger AI models. President Joe Biden signed the CHIPS and Science Act in 2022 to bring chipmaking back to the U.S., and the Biden administration has already invested billions into semiconductor companies including Intel and TSMC to build factories throughout the country. Part of that effort also has to do with countering China’s advancements in chipmaking and AI development.

Advertisement
Previous Slide
Next Slide

6 / 16

What are GPUs & CPUs?

What are GPUs & CPUs?

illustration of a CPU in a computer motherboard with other switches
Core processing unit (CPU) in the motherboard.
Photo: Narumon Bowonkitwanchai (Getty Images)

A GPU is a graphics processing unit, an advanced chip (or semiconductor) that powers the large language models behind AI chatbots like ChatGPT. It was traditionally used to make video games with higher quality visuals.

Advertisement

Then a Ukrainian-Canadian computer scientist, Alex Krizhevsky, showed how using a GPU could power deep learning models a whole lot faster than a CPU — a central processing unit, or the main hardware that powers computers.

CPUs are the “brain” of a computer, carrying out instructions for that computer to work. A CPU is a processor, which reads and interprets software instructions to control the computer’s functions. But a GPU is an accelerator, a piece of hardware designed to advance a specific function of a processor.

Nvidia is the leading GPU designer, with its H100 and H200 chips used in major tech companies’ data centers to power AI software. Other companies are aiming to compete with Nvidia’s accelerators, including Intel with its Gaudi 3 accelerator, and Microsoft’s Azure Maia 100 GPU.

Advertisement
Previous Slide
Next Slide

7 / 16

What is a TPU?

What is a TPU?

A Google video breaks down the ins and outs of its TPU.
A Google video breaks down the ins and outs of its TPU.
Screenshot: Google (Other)

TPU stands for “tensor processing unit.” Google’s chips, unlike those of Microsoft and Nvidia, are TPUs — custom-designed chips made specifically for training large AI models (whereas GPUs were initially made for gaming, not AI).

Advertisement

While CPUs are general-purpose processors and GPUs are an additional processor that run high-end tasks, TPUs are custom-built accelerators to run AI services — making them all the more powerful.

Advertisement
Previous Slide
Next Slide

8 / 16

What is a hallucination?

What is a hallucination?

OpenAI "ChatGPT" AI-generated answer to the question "What can AI offer to humanity?" is seen on a laptop screen
OpenAI’s ChatGPT
Illustration: Leon Neal (Getty Images)

As mentioned before, AI chatbots are capable of a lot of tasks, but they also slip up a lot. When LLMs like ChatGPT make up fake or nonsensical information, that’s called a hallucination.

Advertisement

Chatbots “hallucinate” when they don’t have the necessary training data to answer a question, but still generate a response that looks like a fact. Hallucinations can be caused by different factors such as inaccurate or biased training data and overfitting, which is when an algorithm can’t make predictions or conclusions from other data than what it was trained on.

Hallucinations are currently one of the biggest issues with generative AI models — and they’re not exactly easy to solve for. Because AI models are trained on massive sets of data, it can make it difficult to find specific problems in the data. Sometimes, the data used to train AI models is inaccurate anyway, because it comes from places like Reddit. Although AI models are trained to not answer questions they don’t know the answer to, they sometimes don’t refuse these questions, and instead generate answers that are inaccurate.

Advertisement
Previous Slide
Next Slide

9 / 16

What is training?

What is training?

a little robot is walking on a little treadmill on a desk with a laptop and notebook
Illustration: Westend61 (Getty Images)

Training is the process of teaching an AI model how to make predictions. In this phase, an AI model is fed data to learn how to do a specific task or tasks, and goes through trial and error until it starts producing the desired outputs.

Advertisement
Previous Slide
Next Slide

10 / 16

What is inferencing?

What is inferencing?

illustration of neon blue brain in a tunnel with other blue streaks of light and code
Image: Yuichiro Chino (Getty Images)

Inferencing comes after training, and is the process of a trained AI model making predictions from new data. For example, a self-driving car can be trained to recognize a stop sign on a specific road. Through inferencing, the self-driving car would be able to recognize a stop sign on any road.

Advertisement
Previous Slide
Next Slide

11 / 16

What is retrieval-augmented generation (RAG)?

What is retrieval-augmented generation (RAG)?

illustration of stars shooting out of a brain on a laptop screen against a mint green background
Illustration: Malte Mueller (Getty Images)

Retrieval-augmented generation, or RAG, is a natural language processing (NLP) technique used to improve the accuracy of generative AI models. With RAG, generative large language models are combined with information retrieval systems (like databases and web pages), allowing the models to reference knowledge outside of its original training data and therefore provide more up-to-date answers.

Advertisement

The term was coined in a 2020 paper by a group of researchers from Facebook, University College London, and New York University.

Advertisement
Previous Slide
Next Slide

12 / 16

What are tokens?

What are tokens?

black letters scrambled together, rising against a white background, there are streaks below the letters to illustrate they are moving up
Illustration: carloscastilla (Getty Images)

Text data is broken down into smaller units, called tokens, to be processed by AI models. Tokens can range from being one letter to a whole phrase.

Advertisement
Previous Slide
Next Slide

13 / 16

What are parameters?

What are parameters?

digital illustration of white geometric brain surrounded by colorful dots and lines to show connections
Illustration: da-kuk (Getty Images)

Parameters are the variables a model learns from training data that guide its ability to make predictions. While in the training process, a model adjusts its parameters to close the gap between what it predicts and what the desired outcome is. Therefore, the model learns how to make accurate predictions on new data.

Advertisement
Previous Slide
Next Slide

14 / 16

What is an AI PC?

What is an AI PC?

Businessman hands top view typing in laptop, AI brain hologram, chip with digital circuit board and gears
Illustration: ismagilov (Getty Images)

An AI PC is a personal computer that can handle AI and machine learning tasks. They are built with a CPU, GPU, and NPU that have specific AI capabilities. An NPU, or neural processing unit, is a chip specialized in carrying out AI- and machine learning-based tasks on the computer without having to send that data to be processed in the cloud.

Advertisement
Previous Slide
Next Slide

15 / 16

What is a neural processing unit (NPU)?

What is a neural processing unit (NPU)?

illustration of electronic circuit board with big green chip in the middle with "AI" written in fuchsia letters
Illustration: Jonathan Kitchen (Getty Images)

A neural processing unit (NPU) can run artificial intelligence and machine learning tasks straight on a device, such as an AI PC, meaning the data doesn’t have to be sent to the cloud for processing. 

Advertisement