Artificial quality is simply a heavy and convoluted world. The scientists who enactment successful this tract often trust connected jargon and lingo to explicate what they’re moving on. As a result, we often person to usage those method presumption successful our sum of the artificial quality industry. That’s wherefore we thought it would beryllium adjuvant to enactment unneurotic a glossary with definitions of immoderate of the astir important words and phrases that we usage successful our articles.
We volition regularly update this glossary to adhd caller entries arsenic researchers continually uncover caller methods to propulsion the frontier of artificial quality portion identifying emerging information risks.
AGI
Artificial wide intelligence, oregon AGI, is simply a nebulous term. But it mostly refers to AI that’s much susceptible than the mean quality astatine many, if not most, tasks. OpenAI CEO Sam Altman recently described AGI arsenic the “equivalent of a median quality that you could prosecute arsenic a co-worker.” Meanwhile, OpenAI’s charter defines AGI arsenic “highly autonomous systems that outperform humans astatine astir economically invaluable work.” Google DeepMind’s knowing differs somewhat from these 2 definitions; the laboratory views AGI arsenic “AI that’s astatine slightest arsenic susceptible arsenic humans astatine astir cognitive tasks.” Confused? Not to interest — so are experts astatine the forefront of AI research.
AI agent
An AI cause refers to a instrumentality that uses AI technologies to execute a bid of tasks connected your behalf — beyond what a much basal AI chatbot could bash — specified arsenic filing expenses, booking tickets oregon a array astatine a restaurant, oregon adjacent penning and maintaining code. However, arsenic we’ve explained before, determination are tons of moving pieces successful this emergent space, truthful “AI agent” mightiness mean antithetic things to antithetic people. Infrastructure is besides inactive being built retired to present connected its envisaged capabilities. But the basal conception implies an autonomous strategy that whitethorn gully connected aggregate AI systems to transportation retired multistep tasks.
Chain of thought
Given a elemental question, a quality encephalon tin reply without adjacent reasoning excessively overmuch astir it — things similar “which carnal is taller, a giraffe oregon a cat?” But successful galore cases, you often request a pen and insubstantial to travel up with the close reply due to the fact that determination are intermediary steps. For instance, if a husbandman has chickens and cows, and unneurotic they person 40 heads and 120 legs, you mightiness request to constitute down a elemental equation to travel up with the reply (20 chickens and 20 cows).
In an AI context, chain-of-thought reasoning for ample connection models means breaking down a occupation into smaller, intermediate steps to amended the prime of the extremity result. It usually takes longer to get an answer, but the reply is much apt to beryllium correct, particularly successful a logic oregon coding context. Reasoning models are developed from accepted ample connection models and optimized for chain-of-thought reasoning acknowledgment to reinforcement learning.
(See: Large connection model)
Techcrunch event
San Francisco, CA | October 13-15, 2026
Compute
Although somewhat of a multivalent term, compute mostly refers to the captious computational power that allows AI models to operate. This benignant of processing fuels the AI industry, giving it the quality to bid and deploy its almighty models. The word is often a shorthand for the kinds of hardware that provides the computational powerfulness — things similar GPUs, CPUs, TPUs, and different forms of infrastructure that signifier the bedrock of the modern AI industry.
Deep learning
A subset of self-improving instrumentality learning successful which AI algorithms are designed with a multi-layered, artificial neural web (ANN) structure. This allows them to marque much analyzable correlations compared to simpler instrumentality learning-based systems, specified arsenic linear models oregon determination trees. The operation of heavy learning algorithms draws inspiration from the interconnected pathways of neurons successful the quality brain.
Deep learning AI models are capable to place important characteristics successful information themselves, alternatively than requiring quality engineers to specify these features. The operation besides supports algorithms that tin larn from errors and, done a process of repetition and adjustment, amended their ain outputs. However, heavy learning systems necessitate a batch of information points to output bully results (millions oregon more). They besides typically instrumentality longer to bid compared to simpler instrumentality learning algorithms — truthful improvement costs thin to beryllium higher.
(See: Neural network)
Diffusion
Diffusion is the tech astatine the bosom of galore art-, music-, and text-generating AI models. Inspired by physics, diffusion systems dilatory “destroy” the operation of data — for example, photos, songs, and truthful connected — by adding sound until there’s thing left. In physics, diffusion is spontaneous and irreversible — sweetener diffused successful java can’t beryllium restored to cube form. But diffusion systems successful AI purpose to larn a benignant of “reverse diffusion” process to reconstruct the destroyed data, gaining the quality to retrieve the information from noise.
Distillation
Distillation is simply a method utilized to extract cognition from a ample AI exemplary with a ‘teacher-student’ model. Developers nonstop requests to a teacher exemplary and grounds the outputs. Answers are sometimes compared with a dataset to spot however close they are. These outputs are past utilized to bid the pupil model, which is trained to approximate the teacher’s behavior.
Distillation tin beryllium utilized to make a smaller, much businesslike exemplary based connected a larger exemplary with a minimal distillation loss. This is apt however OpenAI developed GPT-4 Turbo, a faster mentation of GPT-4.
While each AI companies usage distillation internally, it whitethorn person besides been utilized by immoderate AI companies to drawback up with frontier models. Distillation from a rival usually violates the presumption of work of AI API and chat assistants.
Fine-tuning
This refers to the further grooming of an AI exemplary to optimize show for a much circumstantial task oregon country than was antecedently a focal constituent of its grooming — typically by feeding successful new, specialized (i.e., task-oriented) data.
Many AI startups are taking ample connection models arsenic a starting constituent to physique a commercialized merchandise but are vying to amp up inferior for a people assemblage oregon task by supplementing earlier grooming cycles with fine-tuning based connected their ain domain-specific cognition and expertise.
(See: Large connection exemplary [LLM])
GAN
A GAN, oregon Generative Adversarial Network, is simply a benignant of instrumentality learning model that underpins immoderate important developments successful generative AI erstwhile it comes to producing realistic information – including (but not only) deepfake tools. GANs impact the usage of a brace of neural networks, 1 of which draws connected its grooming information to make an output that is passed to the different exemplary to evaluate. This second, discriminator exemplary frankincense plays the relation of a classifier connected the generator’s output – enabling it to amended implicit time.
The GAN operation is acceptable up arsenic a contention (hence “adversarial”) – with the 2 models fundamentally programmed to effort to outdo each other: the generator is trying to get its output past the discriminator, portion the discriminator is moving to spot artificially generated data. This structured contention tin optimize AI outputs to beryllium much realistic without the request for further quality intervention. Though GANs enactment champion for narrower applications (such arsenic producing realistic photos oregon videos), alternatively than wide intent AI.
Hallucination
Hallucination is the AI industry’s preferred word for AI models making worldly up – virtually generating accusation that is incorrect. Obviously, it’s a immense occupation for AI quality.
Hallucinations nutrient GenAI outputs that tin beryllium misleading and could adjacent pb to real-life risks — with perchance unsafe consequences (think of a wellness query that returns harmful aesculapian advice). This is wherefore astir GenAI tools’ tiny people present warns users to verify AI-generated answers, adjacent though specified disclaimers are usually acold little salient than the accusation the tools dispense astatine the interaction of a button.
The occupation of AIs fabricating accusation is thought to originate arsenic a effect of gaps successful grooming data. For wide intent GenAI particularly — besides sometimes known arsenic instauration models — this looks hard to resolve. There is simply not capable information successful beingness to bid AI models to comprehensively resoluteness each the questions we could perchance ask. TL;DR: we haven’t invented God (yet).
Hallucinations are contributing to a propulsion towards progressively specialized and/or vertical AI models — i.e. domain-specific AIs that necessitate narrower expertise – arsenic a mode to trim the likelihood of cognition gaps and shrink disinformation risks.
Inference
Inference is the process of moving an AI model. It’s mounting a exemplary escaped to marque predictions oregon gully conclusions from antecedently seen data. To beryllium clear, inference can’t hap without training; a exemplary indispensable larn patterns successful a acceptable of information earlier it tin efficaciously extrapolate from this grooming data.
Many types of hardware tin execute inference, ranging from smartphone processors to beefy GPUs to custom-designed AI accelerators. But not each of them tin tally models arsenic well. Very ample models would instrumentality ages to marque predictions on, say, a laptop versus a unreality server with high-end AI chips.
[See: Training]
Large connection exemplary (LLM)
Large connection models, oregon LLMs, are the AI models utilized by fashionable AI assistants, specified arsenic ChatGPT, Claude, Google’s Gemini, Meta’s AI Llama, Microsoft Copilot, oregon Mistral’s Le Chat. When you chat with an AI assistant, you interact with a ample connection exemplary that processes your petition straight oregon with the assistance of antithetic disposable tools, specified arsenic web browsing oregon codification interpreters.
AI assistants and LLMs tin person antithetic names. For instance, GPT is OpenAI’s ample connection exemplary and ChatGPT is the AI adjunct product.
LLMs are heavy neural networks made of billions of numerical parameters (or weights, spot below) that larn the relationships betwixt words and phrases and make a practice of language, a benignant of multidimensional representation of words.
These models are created from encoding the patterns they find successful billions of books, articles, and transcripts. When you punctual an LLM, the exemplary generates the astir apt signifier that fits the prompt. It past evaluates the astir probable adjacent connection aft the past 1 based connected what was said before. Repeat, repeat, and repeat.
(See: Neural network)
Memory Cache
Memory cache refers to an important process that boosts inference (which is the process by which AI works to make a effect to a user’s query). In essence, caching is an optimization technique, designed to marque inference much efficient. AI is evidently driven by high-octane mathematical calculations and each clip those calculations are made, they usage up much power. Caching is designed to chopped down connected the fig of calculations a exemplary mightiness person to tally by redeeming peculiar calculations for aboriginal idiosyncratic queries and operations. There are antithetic kinds of representation caching, though 1 of the much well-known is KV (or cardinal value) caching. KV caching works successful transformer-based models, and increases efficiency, driving faster results by reducing the magnitude of clip (and algorithmic labor) it takes to make answers to idiosyncratic questions.
(See: Inference)
Neural network
A neural web refers to the multi-layered algorithmic operation that underpins heavy learning — and, much broadly, the full roar successful generative AI tools pursuing the emergence of ample connection models.
Although the thought of taking inspiration from the densely interconnected pathways of the quality encephalon arsenic a plan operation for information processing algorithms dates each the mode backmost to the 1940s, it was the overmuch much caller emergence of graphical processing hardware (GPUs) — via the video crippled manufacture — that truly unlocked the powerfulness of this theory. These chips proved good suited to grooming algorithms with galore much layers than was imaginable successful earlier epochs — enabling neural network-based AI systems to execute acold amended show crossed galore domains, including dependable recognition, autonomous navigation, and cause discovery.
(See: Large connection exemplary [LLM])
RAMageddon
RAMageddon is the amusive caller word for a not-so-fun inclination that is sweeping the tech industry: an ever-increasing shortage of random entree memory, oregon RAM chips, which powerfulness beauteous overmuch each the tech products we usage successful our regular lives. As the AI manufacture has blossomed, the biggest tech companies and AI labs — each vying to person the astir almighty and businesslike AI — are buying truthful overmuch RAM to powerfulness their information centers that there’s not overmuch near for the remainder of us. And that proviso bottleneck means that what’s near is getting much and much expensive.
That includes industries similar gaming (where large companies person had to raise prices connected consoles due to the fact that it’s harder to find representation chips for their devices), user electronics (where representation shortage could origin the biggest dip successful smartphone shipments successful much than a decade), and wide endeavor computing (because those companies can’t get capable RAM for their ain information centers). The surge successful prices is lone expected to halt aft the dreaded shortage ends but, unfortunately, there’s not truly overmuch of a sign that’s going to hap anytime soon.
Training
Developing instrumentality learning AIs involves a process known arsenic training. In elemental terms, this refers to information being fed successful in bid that the exemplary tin larn from patterns and make utile outputs.
Things tin get a spot philosophical astatine this constituent successful the AI stack — since, pre-training, the mathematical operation that’s utilized arsenic the starting constituent for processing a learning strategy is conscionable a clump of layers and random numbers. It’s lone done grooming that the AI exemplary truly takes shape. Essentially, it’s the process of the strategy responding to characteristics successful the information that enables it to accommodate outputs towards a sought-for extremity — whether that’s identifying images of cats oregon producing a haiku connected demand.
It’s important to enactment that not each AI requires training. Rules-based AIs that are programmed to travel manually predefined instructions — for example, specified arsenic linear chatbots — don’t request to acquisition training. However, specified AI systems are apt to beryllium much constrained than (well-trained) self-learning systems.
Still, grooming tin beryllium costly due to the fact that it requires tons of inputs — and, typically, the volumes of inputs required for specified models person been trending upwards.
Hybrid approaches tin sometimes beryllium utilized to shortcut exemplary improvement and assistance negociate costs. Such arsenic doing data-driven fine-tuning of a rules-based AI — meaning improvement requires little data, compute, energy, and algorithmic complexity than if the developer had started gathering from scratch.
[See: Inference]
Tokens
When it comes to human-machine communication, determination are immoderate evident challenges. People pass utilizing quality language, portion AI programs execute tasks and respond to queries done analyzable algorithmic processes that are informed by data. In their simplest definition, tokens correspond the basal gathering blocks of human-AI communication, successful that they are discrete segments of information that person either been processed oregon produced by an LLM.
Tokens are created via a process known arsenic “tokenization,” which breaks down earthy information and refines it into chiseled units that are digestible to an LLM. Similar to however a bundle compiler translates quality connection into binary codification that a machine tin digest, tokenization interprets quality connection for an AI programme via their idiosyncratic queries truthful that it tin hole a response.
There are respective antithetic kinds of tokens — including input tokens (the benignant that indispensable beryllium generated successful effect to a quality user’s query), output tokens (the benignant that are generated arsenic the LLM responds to the human’s request), and reasoning tokens, which impact longer, much intensive tasks and processes that hap arsenic portion of a idiosyncratic request.
With endeavor AI, token usage besides determines costs. Since tokens are equivalent to the magnitude of information being processed by a model, they person besides go the means by which the AI manufacture monetizes its services. Most AI companies complaint for LLM usage connected a per-token-basis. Thus, the much tokens a concern burns arsenic it uses an AI programme (ChatGPT, for example), the much wealth it volition person to wage its AI work supplier (OpenAI).
Transfer learning
A method wherever a antecedently trained AI exemplary is utilized arsenic the starting constituent for processing a caller exemplary for a antithetic but typically related task – allowing cognition gained successful erstwhile grooming cycles to beryllium reapplied.
Transfer learning tin thrust ratio savings by shortcutting exemplary development. It tin besides beryllium utile erstwhile information for the task that the exemplary is being developed for is somewhat limited. But it’s important to enactment that the attack has limitations. Models that trust connected transportation learning to summation generalized capabilities volition apt necessitate grooming connected further information successful bid to execute good successful their domain of focus
(See: Fine tuning)
Weights
Weights are halfway to AI training, arsenic they find however overmuch value (or weight) is fixed to antithetic features (or input variables) successful the information utilized for grooming the strategy — thereby shaping the AI model’s output.
Put different way, weights are numerical parameters that specify what’s astir salient successful a dataset for the fixed grooming task. They execute their relation by applying multiplication to inputs. Model grooming typically begins with weights that are randomly assigned, but arsenic the process unfolds, the weights set arsenic the exemplary seeks to get astatine an output that much intimately matches the target.
For example, an AI exemplary for predicting lodging prices that’s trained connected humanities existent property information for a people determination could see weights for features specified arsenic the fig of bedrooms and bathrooms, whether a spot is detached oregon semi-detached, whether it has parking, a garage, and truthful on.
Ultimately, the weights the exemplary attaches to each of these inputs bespeak however overmuch they power the worth of a property, based connected the fixed dataset.
This nonfiction is updated regularly with caller information.















English (US) ·