From LLMs to hallucinations, here’s a simple guide to common AI terms - BERITAJA

Albert Michael By: Albert Michael - Sunday, 12 April 2026 22:07:08 • 16 min read
From LLMs to hallucinations, here’s a simple guide to common AI terms - BERITAJA

From LLMs to hallucinations, here’s a simple guide to common AI terms - BERITAJA is one of the most discussed topics today. In this article, you will find a clear explanation, key facts, and the latest updates related to this topic, presented in a concise and easy-to-understand way. Read more news on Beritaja.

Artificial intelligence is simply a heavy and convoluted world. The scientists who activity successful this section often trust connected motto and lingo to explicate what they’re moving on. As a result, we often person to usage those method position successful our sum of the artificial intelligence industry. That’s why we thought it would beryllium adjuvant to put together a glossary pinch definitions of immoderate of the about important words and phrases that we usage successful our articles.

We will regularly update this glossary to adhd caller entries arsenic researchers continually uncover caller methods to push the frontier of artificial intelligence while identifying emerging information risks.


AGI

Artificial wide intelligence, aliases AGI, is simply a nebulous term. But it mostly refers to AI that’s much could than the mean quality astatine many, if not most, tasks. OpenAI CEO Sam Altman recently described AGI arsenic the “equivalent of a median quality that you could prosecute arsenic a co-worker.” Meanwhile, OpenAI’s charter defines AGI arsenic “highly autonomous systems that outperform humans astatine about economically valuable work.” Google DeepMind’s knowing differs somewhat from these 2 definitions; the laboratory views AGI arsenic “AI that’s astatine slightest arsenic could arsenic humans astatine about cognitive tasks.” Confused? Not to interest — so are experts astatine the forefront of AI research.

AI agent

An AI supplier refers to a instrumentality that uses AI technologies to execute a bid of tasks connected your behalf — beyond what a much basal AI chatbot could do — specified arsenic filing expenses, booking tickets aliases a array astatine a restaurant, aliases moreover penning and maintaining code. However, arsenic we’ve explained before, location are tons of moving pieces successful this emergent space, truthful “AI agent” mightiness mean different things to different people. Infrastructure is besides still being built retired to present connected its envisaged capabilities. But the basal conception implies an autonomous strategy that whitethorn tie connected aggregate AI systems to transportation retired multistep tasks.

Chain of thought

Given a elemental question, a quality encephalon could reply without moreover reasoning excessively overmuch about it — things for illustration “which animal is taller, a giraffe aliases a cat?” But successful galore cases, you often request a pen and insubstantial to travel up pinch the correct reply because location are intermediary steps. For instance, if a husbandman has chickens and cows, and together they person 40 heads and 120 legs, you mightiness request to constitute down a elemental equation to travel up pinch the reply (20 chickens and 20 cows).

In an AI context, chain-of-thought reasoning for ample connection models intends breaking down a problem into smaller, intermediate steps to amended the value of the extremity result. It usually takes longer to get an answer, but the reply is much apt to beryllium correct, particularly successful a logic aliases coding context. Reasoning models are developed from accepted ample connection models and optimized for chain-of-thought reasoning acknowledgment to reinforcement learning.

(See: Large connection model)

Techcrunch event

San Francisco, CA | October 13-15, 2026

Compute

Although somewhat of a multivalent term, compute mostly refers to the captious computational power that allows AI models to operate. This type of processing fuels the AI industry, giving it the expertise to train and deploy its powerful models. The word is often a shorthand for the kinds of hardware that provides the computational powerfulness — things for illustration GPUs, CPUs, TPUs, and different forms of infrastructure that shape the bedrock of the modern AI industry.

Deep learning

A subset of self-improving instrumentality learning successful which AI algorithms are designed pinch a multi-layered, artificial neural web (ANN) structure. This allows them to make much analyzable correlations compared to simpler instrumentality learning-based systems, specified arsenic linear models aliases determination trees. The building of heavy learning algorithms draws inspiration from the interconnected pathways of neurons successful the quality brain.

Deep learning AI models are capable to place important characteristics successful information themselves, alternatively than requiring quality engineers to specify these features. The building besides supports algorithms that could study from errors and, done a process of repetition and adjustment, amended their ain outputs. However, heavy learning systems require a batch of information points to output bully results (millions aliases more). They besides typically return longer to train compared to simpler instrumentality learning algorithms — truthful improvement costs thin to beryllium higher.

(See: Neural network)

Diffusion

Diffusion is the tech astatine the bosom of galore art-, music-, and text-generating AI models. Inspired by physics, diffusion systems slow “destroy” the building of data — for example, photos, songs, and truthful connected — by adding sound until there’s thing left. In physics, diffusion is spontaneous and irreversible — sweetener diffused successful java can’t beryllium restored to cube form. But diffusion systems successful AI purpose to study a benignant of “reverse diffusion” process to reconstruct the destroyed data, gaining the expertise to retrieve the information from noise.

Distillation

Distillation is simply a method utilized to extract knowledge from a ample AI exemplary pinch a ‘teacher-student’ model. Developers nonstop requests to a coach exemplary and grounds the outputs. Answers are sometimes compared pinch a dataset to spot really meticulous they are. These outputs are past utilized to train the student model, which is trained to approximate the teacher’s behavior.

Distillation could beryllium utilized to create a smaller, much businesslike exemplary based connected a larger exemplary pinch a minimal distillation loss. This is apt really OpenAI developed GPT-4 Turbo, a faster type of GPT-4.

While each AI companies usage distillation internally, it whitethorn person besides been utilized by immoderate AI companies to drawback up pinch frontier models. Distillation from a competitor usually violates the position of work of AI API and chat assistants.

Fine-tuning

This refers to the further training of an AI exemplary to optimize capacity for a much circumstantial task aliases area than was antecedently a focal constituent of its training — typically by feeding successful new, specialized (i.e., task-oriented) data. 

Many AI startups are taking ample connection models arsenic a starting constituent to build a commercialized merchandise but are vying to amp up inferior for a target assemblage aliases task by supplementing earlier training cycles pinch fine-tuning based connected their ain domain-specific knowledge and expertise.

(See: Large connection exemplary [LLM])

GAN

A GAN, aliases Generative Adversarial Network, is simply a type of instrumentality learning model that underpins immoderate important developments successful generative AI erstwhile it comes to producing realistic information – including (but not only) deepfake tools. GANs impact the usage of a brace of neural networks, 1 of which draws connected its training information to make an output that is passed to the different exemplary to evaluate. This second, discriminator exemplary frankincense plays the domiciled of a classifier connected the generator’s output – enabling it to amended complete time. 

The GAN building is group up arsenic a title (hence “adversarial”) – pinch the 2 models fundamentally programmed to effort to outdo each other: the generator is trying to get its output past the discriminator, while the discriminator is moving to spot artificially generated data. This system title could optimize AI outputs to beryllium much realistic without the request for further quality intervention. Though GANs activity champion for narrower applications (such arsenic producing realistic photos aliases videos), alternatively than wide intent AI.

Hallucination

Hallucination is the AI industry’s preferred word for AI models making worldly up – virtually generating accusation that is incorrect. Obviously, it’s a immense problem for AI quality. 

Hallucinations nutrient GenAI outputs that could beryllium misleading and could moreover lead to real-life risks — pinch perchance vulnerable consequences (think of a wellness query that returns harmful aesculapian advice). This is why about GenAI tools’ mini people now warns users to verify AI-generated answers, moreover though specified disclaimers are usually acold little salient than the accusation the devices dispense astatine the touch of a button.

The problem of AIs fabricating accusation is thought to originate arsenic a consequence of gaps successful training data. For wide intent GenAI particularly — besides sometimes known arsenic instauration models — this looks difficult to resolve. There is simply not capable information successful beingness to train AI models to comprehensively resoluteness each the questions we could perchance ask. TL;DR: we haven’t invented God (yet). 

Hallucinations are contributing to a push towards progressively specialized and/or vertical AI models — i.e. domain-specific AIs that require narrower expertise – arsenic a measurement to trim the likelihood of knowledge gaps and shrink disinformation risks.

Inference

Inference is the process of moving an AI model. It’s mounting a exemplary loose to make predictions aliases tie conclusions from antecedently seen data. To beryllium clear, conclusion can’t hap without training; a exemplary must study patterns successful a group of information earlier it could efficaciously extrapolate from this training data.

Many types of hardware could execute inference, ranging from smartphone processors to beefy GPUs to custom-designed AI accelerators. But not each of them could tally models arsenic well. Very ample models would return ages to make predictions on, say, a laptop versus a unreality server pinch high-end AI chips.

[See: Training]

Large connection exemplary (LLM)

Large connection models, aliases LLMs, are the AI models utilized by celebrated AI assistants, specified arsenic ChatGPT, Claude, Google’s Gemini, Meta’s AI Llama, Microsoft Copilot, aliases Mistral’s Le Chat. When you chat pinch an AI assistant, you interact pinch a ample connection exemplary that processes your petition straight aliases pinch the thief of different disposable tools, specified arsenic web browsing aliases codification interpreters.

AI assistants and LLMs could person different names. For instance, GPT is OpenAI’s ample connection exemplary and ChatGPT is the AI adjunct product.

LLMs are heavy neural networks made of billions of numerical parameters (or weights, spot below) that study the relationships betwixt words and phrases and create a practice of language, a benignant of multidimensional representation of words.

These models are created from encoding the patterns they find successful billions of books, articles, and transcripts. When you punctual an LLM, the exemplary generates the about apt shape that fits the prompt. It past evaluates the about probable adjacent connection aft the past 1 based connected what was said before. Repeat, repeat, and repeat.

(See: Neural network)

Memory Cache

Memory cache refers to an important process that boosts conclusion (which is the process by which AI useful to make a consequence to a user’s query). In essence, caching is an optimization technique, designed to make conclusion much efficient. AI is evidently driven by high-octane mathematical calculations and each clip those calculations are made, they usage up much power. Caching is designed to trim down connected the number of calculations a exemplary mightiness person to tally by redeeming peculiar calculations for early personification queries and operations. There are different kinds of representation caching, though 1 of the much well-known is KV (or cardinal value) caching. KV caching useful successful transformer-based models, and increases efficiency, driving faster results by reducing the magnitude of clip (and algorithmic labor) it takes to make answers to personification questions.   

(See: Inference)  

Neural network

A neural web refers to the multi-layered algorithmic building that underpins heavy learning — and, much broadly, the full roar successful generative AI devices pursuing the emergence of ample connection models. 

Although the thought of taking inspiration from the densely interconnected pathways of the quality encephalon arsenic a creation building for information processing algorithms dates each the measurement backmost to the 1940s, it was the overmuch much caller emergence of graphical processing hardware (GPUs) — via the video crippled manufacture — that really unlocked the powerfulness of this theory. These chips proved good suited to training algorithms pinch galore much layers than was imaginable successful earlier epochs — enabling neural network-based AI systems to execute acold amended capacity crossed galore domains, including sound recognition, autonomous navigation, and supplier discovery.

(See: Large connection exemplary [LLM])

RAMageddon

RAMageddon is the nosy caller word for a not-so-fun inclination that is sweeping the tech industry: an ever-increasing shortage of random entree memory, aliases RAM chips, which powerfulness beautiful overmuch each the tech products we usage successful our regular lives. As the AI manufacture has blossomed, the biggest tech companies and AI labs — each vying to person the about powerful and businesslike AI — are buying truthful overmuch RAM to powerfulness their information centers that there’s not overmuch near for the remainder of us. And that proviso bottleneck intends that what’s near is getting much and much expensive.

That includes industries for illustration gaming (where awesome companies person had to raise prices connected consoles because it’s harder to find representation chips for their devices), user electronics (where representation shortage could origin the biggest dip successful smartphone shipments successful much than a decade), and wide endeavor computing (because those companies can’t get capable RAM for their ain information centers). The surge successful prices is only expected to extremity aft the dreaded shortage ends but, unfortunately, there’s not really overmuch of a sign that’s going to hap anytime soon.  

Training

Developing instrumentality learning AIs involves a process known arsenic training. In elemental terms, this refers to information being fed successful in bid that the exemplary could study from patterns and make useful outputs.

Things could get a spot philosophical astatine this constituent successful the AI stack — since, pre-training, the mathematical building that’s utilized arsenic the starting constituent for processing a learning strategy is conscionable a bunch of layers and random numbers. It’s only done training that the AI exemplary really takes shape. Essentially, it’s the process of the strategy responding to characteristics successful the information that enables it to accommodate outputs towards a sought-for extremity — whether that’s identifying images of cats aliases producing a haiku connected demand.

It’s important to statement that not each AI requires training. Rules-based AIs that are programmed to travel manually predefined instructions — for example, specified arsenic linear chatbots — don’t request to acquisition training. However, specified AI systems are apt to beryllium much constrained than (well-trained) self-learning systems.

Still, training could beryllium costly because it requires tons of inputs — and, typically, the volumes of inputs required for specified models person been trending upwards.

Hybrid approaches could sometimes beryllium utilized to shortcut exemplary improvement and thief negociate costs. Such arsenic doing data-driven fine-tuning of a rules-based AI — meaning improvement requires little data, compute, energy, and algorithmic complexity than if the developer had started building from scratch.

[See: Inference]

Tokens

When it comes to human-machine communication, location are immoderate evident challenges. People pass utilizing quality language, while AI programs execute tasks and respond to queries done analyzable algorithmic processes that are informed by data. In their simplest definition, tokens correspond the basal building blocks of human-AI communication, successful that they are discrete segments of information that person either been processed aliases produced by an LLM. 

Tokens are created via a process known arsenic “tokenization,” which breaks down earthy information and refines it into chopped units that are digestible to an LLM. Similar to really a package compiler translates quality connection into binary codification that a machine could digest, tokenization interprets quality connection for an AI programme via their personification queries truthful that it could hole a response. 

There are respective different kinds of tokens — including input tokens (the benignant that must beryllium generated successful consequence to a quality user’s query), output tokens (the benignant that are generated arsenic the LLM responds to the human’s request), and reasoning tokens, which impact longer, much intensive tasks and processes that hap arsenic portion of a personification request. 

With endeavor AI, token usage besides determines costs. Since tokens are balanced to the magnitude of information being processed by a model, they person besides go the intends by which the AI manufacture monetizes its services. Most AI companies complaint for LLM usage connected a per-token-basis. Thus, the much tokens a business burns arsenic it uses an AI programme (ChatGPT, for example), the much money it will person to salary its AI work supplier (OpenAI). 

Transfer learning

A method wherever a antecedently trained AI exemplary is utilized arsenic the starting constituent for processing a caller exemplary for a different but typically related task – allowing knowledge gained successful erstwhile training cycles to beryllium reapplied. 

Transfer learning could thrust ratio savings by shortcutting exemplary development. It could besides beryllium useful erstwhile information for the task that the exemplary is being developed for is somewhat limited. But it’s important to statement that the attack has limitations. Models that trust connected transportation learning to summation generalized capabilities will apt require training connected further information successful bid to execute good successful their domain of focus

(See: Fine tuning)

Weights

Weights are halfway to AI training, arsenic they find really overmuch value (or weight) is fixed to different features (or input variables) successful the information utilized for training the strategy — thereby shaping the AI model’s output. 

Put different way, weights are numerical parameters that specify what’s about salient successful a dataset for the fixed training task. They execute their usability by applying multiplication to inputs. Model training typically originates pinch weights that are randomly assigned, but arsenic the process unfolds, the weights set arsenic the exemplary seeks to get astatine an output that much intimately matches the target.

For example, an AI exemplary for predicting lodging prices that’s trained connected humanities existent property information for a target location could see weights for features specified arsenic the number of bedrooms and bathrooms, whether a spot is detached aliases semi-detached, whether it has parking, a garage, and truthful on. 

Ultimately, the weights the exemplary attaches to each of these inputs bespeak really overmuch they power the worth of a property, based connected the fixed dataset.

This article is updated regularly pinch caller information.

This article discusses From LLMs to hallucinations, here’s a simple guide to common AI terms - BERITAJA in detail, including key facts, recent developments, and important insights that readers are actively searching for online.