Geoffrey Hinton
(164 articles)![Convolution](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fconvolution.webp&w=3840&q=75)
Convolution
Mathematical operation used in signal processing and image processing to combine two functions, resulting in a third function that represents how one function modifies the other.
Generality: 870
![Tensor](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Ftensor.webp&w=3840&q=75)
Tensor
Multi-dimensional array used in mathematics and computer science, serving as a fundamental data structure in neural networks for representing data and parameters.
Generality: 920
![Parameterized](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fparameterized.webp&w=3840&q=75)
Parameterized
Model or function in AI that utilizes parameters to make predictions or decisions.
Generality: 796
![Loss Optimization](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Floss-optimization.webp&w=3840&q=75)
Loss Optimization
Process of adjusting a model's parameters to minimize the difference between the predicted outputs and the actual outputs, measured by a loss function.
Generality: 886
![Quantization](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fquantization.webp&w=3840&q=75)
Quantization
Process of reducing the precision of the weights and activations in neural network models to decrease their memory and computational requirements.
Generality: 673
![ANN (Artificial Neural Networks)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fann-artificial-neural-networks.webp&w=3840&q=75)
ANN
Artificial Neural Networks
Artificial Neural Networks
Computing systems inspired by the biological neural networks that constitute animal brains, designed to progressively improve their performance on tasks by considering examples.
Generality: 875
![Neural Network](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fneural-network.webp&w=3840&q=75)
Neural Network
Computing system designed to simulate the way human brains analyze and process information, using a network of interconnected nodes that work together to solve specific problems.
Generality: 932
![Connectionist AI](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fconnectionist-ai.webp&w=3840&q=75)
Connectionist AI
Set of computational models in AI that simulate the human brain's network of neurons to process information and learn from data.
Generality: 900
![Next Word Prediction](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fnext-word-prediction.webp&w=3840&q=75)
Next Word Prediction
Enables language models to predict the most probable subsequent word in a text sequence using generative AI techniques.
Generality: 780
![NLP (Natural Language Processing)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fnlp-natural-language-processing.webp&w=3840&q=75)
NLP
Natural Language Processing
Natural Language Processing
Field of AI that focuses on the interaction between computers and humans through natural language.
Generality: 931
![Natural Language Problem](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fnatural-language-problem.webp&w=3840&q=75)
Natural Language Problem
Challenges encountered in understanding, processing, or generating human language using computational methods.
Generality: 875
![Speech Processing](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fspeech-processing.webp&w=3840&q=75)
Speech Processing
Technology that enables computers to recognize, interpret, and generate human speech.
Generality: 870
![ASR (Automatic Speech Recognition)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fasr-automatic-speech-recognition.webp&w=3840&q=75)
ASR
Automatic Speech Recognition
Automatic Speech Recognition
Translates spoken language into written text, enabling computers to understand and process human speech.
Generality: 830
![Generalization](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fgeneralization.webp&w=3840&q=75)
Generalization
Ability of a ML model to perform well on new, unseen data that was not included in the training set.
Generality: 891
![Motor Learning](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fmotor-learning.webp&w=3840&q=75)
Motor Learning
Process by which robots or AI systems acquire, refine, and optimize motor skills through experience and practice.
Generality: 675
![Supervision](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fsupervision.webp&w=3840&q=75)
Supervision
Use of labeled data to train ML models, guiding the learning process by providing input-output pairs.
Generality: 890
![Training](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Ftraining.webp&w=3840&q=75)
Training
Process of teaching a ML model to make accurate predictions or decisions, by adjusting its parameters based on data.
Generality: 940
![Human-Level AI](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fhuman-level-ai.webp&w=3840&q=75)
Human-Level AI
AI systems that can perform any intellectual task with the same proficiency as a human being.
Generality: 945
![AGI (Artificial General Intelligence)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fagi-artificial-general-intelligence.webp&w=3840&q=75)
AGI
Artificial General Intelligence
Artificial General Intelligence
AI capable of understanding, learning, and applying knowledge across a wide range of tasks, matching or surpassing human intelligence.
Generality: 905
![Unsupervised Learning](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Funsupervised-learning.webp&w=3840&q=75)
Unsupervised Learning
Type of ML where algorithms learn patterns from untagged data, without any guidance on what outcomes to predict.
Generality: 905
![Artificial Neuron](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fartificial-neuron.webp&w=3840&q=75)
Artificial Neuron
Computational models inspired by biological neurons, serving as the foundational units of artificial neural networks to process input and output signals.
Generality: 825
![MCP neuron](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fmcp-neuron.webp&w=3840&q=75)
MCP neuron
Early computational model of a biological neuron forming the basis for artificial neural networks.
Generality: 500
![ML (Machine Learning)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fml-machine-learning.webp&w=3840&q=75)
ML
Machine Learning
Machine Learning
Development of algorithms and statistical models that enable computers to perform tasks without being explicitly programmed for each one.
Generality: 965
![Supervised Learning](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fsupervised-learning.webp&w=3840&q=75)
Supervised Learning
ML approach where models are trained on labeled data to predict outcomes or classify data into categories.
Generality: 882
![Memory Systems](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fmemory-systems.webp&w=3840&q=75)
Memory Systems
Mechanisms and structures designed to store, manage, and recall information, enabling machines to learn from past experiences and perform complex tasks.
Generality: 790
![Internal Representation](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Finternal-representation.webp&w=3840&q=75)
Internal Representation
The way information is structured and stored within an AI system, enabling the system to process, reason, or make decisions.
Generality: 845
![Image Recognition](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fimage-recognition.webp&w=3840&q=75)
Image Recognition
Ability of AI to identify objects, places, people, writing, and actions in images.
Generality: 854
![Linear Separability](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Flinear-separability.webp&w=3840&q=75)
Linear Separability
The ability of a dataset to be perfectly separated into two classes using a straight line in two dimensions or a hyperplane in higher dimensions.
Generality: 500
![Pattern Recognition](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fpattern-recognition.webp&w=3840&q=75)
Pattern Recognition
The identification and classification of patterns in data using computational algorithms, essential for enabling machines to interpret, learn from, and make decisions based on complex datasets.
Generality: 825
![Feed Forward](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Ffeed-forward.webp&w=3840&q=75)
Feed Forward
Essential structure of an artificial neural network that directs data or information from the input layer towards the output layer without looping back.
Generality: 860
![Function Approximation](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Ffunction-approximation.webp&w=3840&q=75)
Function Approximation
Method used in AI to estimate complex functions using simpler, computationally efficient models.
Generality: 810
![Invariance](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Finvariance.webp&w=3840&q=75)
Invariance
Property of a model or algorithm that ensures its output remains unchanged when specific transformations are applied to the input data.
Generality: 830
![Inference](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Finference.webp&w=3840&q=75)
Inference
Process by which a trained neural network applies learned patterns to new, unseen data to make predictions or decisions.
Generality: 861
![Perceptron Convergence](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fperceptron-convergence.webp&w=3840&q=75)
Perceptron Convergence
A phenomena where a perceptron algorithm effectively stabilizes, ensuring that it can find a solution for linearly separable datasets after a finite number of iterations.
Generality: 500
![Regularization](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fregularization.webp&w=3840&q=75)
Regularization
Technique used in machine learning to reduce model overfitting by adding a penalty to the loss function based on the complexity of the model.
Generality: 845
![NLU (Natural Language Understanding)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fnlu-natural-language-understanding.webp&w=3840&q=75)
NLU
Natural Language Understanding
Natural Language Understanding
Subfield of NLP focused on enabling machines to understand and interpret human language in a way that is both meaningful and contextually relevant.
Generality: 894
![Overfitting](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Foverfitting.webp&w=3840&q=75)
Overfitting
When a ML model learns the detail and noise in the training data to the extent that it negatively impacts the performance of the model on new data.
Generality: 890
![CNN (Convolutional Neural Network)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fcnn-convolutional-neural-network.webp&w=3840&q=75)
CNN
Convolutional Neural Network
Convolutional Neural Network
Deep learning algorithm that can capture spatial hierarchies in data, particularly useful for image and video recognition tasks.
Generality: 916
![Local Weight Sharing](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Flocal-weight-sharing.webp&w=3840&q=75)
Local Weight Sharing
Technique where the same weights are used across different positions in an input, enhancing the network's ability to recognize patterns irrespective of their spatial location.
Generality: 690
![Statistical AI](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fstatistical-ai.webp&w=3840&q=75)
Statistical AI
Utilizes statistical methods to analyze data and make probabilistic inferences, aimed at emulating aspects of human intelligence through quantitative models.
Generality: 890
![Generative](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fgenerative.webp&w=3840&q=75)
Generative
Subset of AI technologies capable of generating new content, ideas, or data that mimic human-like outputs.
Generality: 840
![Speech-to-Text Model](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fspeech-to-text-model.webp&w=3840&q=75)
Speech-to-Text Model
A computational model designed to convert spoken language into written text using AI and linguistic pattern recognition.
Generality: 805
![Universal Learning Algorithms](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Funiversal-learning-algorithms.webp&w=3840&q=75)
Universal Learning Algorithms
Theoretical frameworks aimed at creating systems capable of learning any task to human-level competency, leveraging principles that could allow for generalization across diverse domains.
Generality: 840
![Learnability](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Flearnability.webp&w=3840&q=75)
Learnability
Capacity of an algorithm or model to effectively learn from data, often measured by how well it can generalize from training data to unseen data.
Generality: 847
![Program Induction](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fprogram-induction.webp&w=3840&q=75)
Program Induction
A process in AI where computers generate, or 'induce', programs based on provided data and specific output criteria.
Generality: 785
![Dualism](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fdualism.webp&w=3840&q=75)
Dualism
Theory or concept that emphasizes the division between symbolic (classical) AI and sub-symbolic (connectionist) AI.
Generality: 830
![Boltzmann Machine](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fboltzmann-machine.webp&w=3840&q=75)
Boltzmann Machine
Stochastic recurrent neural network used to learn and represent complex probability distributions over binary variables.
Generality: 790
![EBM (Energy-Based Model)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Febm-energy-based-model.webp&w=3840&q=75)
EBM
Energy-Based Model
Energy-Based Model
Class of deep learning models that learn to associate lower energy levels with more probable configurations of the input data.
Generality: 625
![RBMs (Restricted Boltzmann Machines)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Frbms-restricted-boltzmann-machines.webp&w=3840&q=75)
RBMs
Restricted Boltzmann Machines
Restricted Boltzmann Machines
Type of generative stochastic artificial neural network that can learn a probability distribution over its set of inputs.
Generality: 770
![Saturating Non-Linearities](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fsaturating-non-linearities.webp&w=3840&q=75)
Saturating Non-Linearities
Activation functions in neural networks that reach a point where their output changes very little, or not at all, in response to large input values.
Generality: 575
![MLP (Multilayer Perceptron)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fmlp-multilayer-perceptron.webp&w=3840&q=75)
MLP
Multilayer Perceptron
Multilayer Perceptron
Type of artificial neural network comprised of multiple layers of neurons, with each layer fully connected to the next, commonly used for tasks involving classification and regression.
Generality: 775
![DL (Deep Learning)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fdl-deep-learning.webp&w=3840&q=75)
DL
Deep Learning
Deep Learning
Subset of machine learning that involves neural networks with many layers, enabling the modeling of complex patterns in data.
Generality: 905
![Backpropagation](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fbackpropagation.webp&w=3840&q=75)
Backpropagation
Algorithm used for training artificial neural networks, crucial for optimizing the weights to minimize error between predicted and actual outcomes.
Generality: 890
![Inductive Bias](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Finductive-bias.webp&w=3840&q=75)
Inductive Bias
Assumptions integrated into a learning algorithm to enable it to generalize from specific instances to broader patterns or concepts.
Generality: 827
![DNN (Deep Neural Networks)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fdnn-deep-neural-networks.webp&w=3840&q=75)
DNN
Deep Neural Networks
Deep Neural Networks
Advanced neural network architectures with multiple layers that enable complex pattern recognition and learning from large amounts of data.
Generality: 916
![Subsymbolic AI](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fsubsymbolic-ai.webp&w=3840&q=75)
Subsymbolic AI
AI approaches that do not use explicit symbolic representation of knowledge but instead rely on distributed, often neural network-based methods to process and learn from data.
Generality: 900
![RNN (Recurrent Neural Network)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Frnn-recurrent-neural-network.webp&w=3840&q=75)
RNN
Recurrent Neural Network
Recurrent Neural Network
Class of neural networks where connections between nodes form a directed graph along a temporal sequence, enabling them to exhibit temporal dynamic behavior for a sequence of inputs.
Generality: 892
![Hidden Layer](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fhidden-layer.webp&w=3840&q=75)
Hidden Layer
Layer of neurons in an artificial neural network that processes inputs from the previous layer, transforming the data before passing it on to the next layer, without direct exposure to the input or output data.
Generality: 861
![Forward Propagation](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fforward-propagation.webp&w=3840&q=75)
Forward Propagation
Process in a neural network where input data is passed through layers of the network to generate output.
Generality: 830
![Feature Extraction](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Ffeature-extraction.webp&w=3840&q=75)
Feature Extraction
Process of transforming raw data into a set of features that are more meaningful and informative for a specific task, such as classification or prediction.
Generality: 880
![Function Approximator](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Ffunction-approximator.webp&w=3840&q=75)
Function Approximator
Computational model used to estimate a target function that is generally complex or unknown, often applied in machine learning and control systems.
Generality: 806
![Weight Initialization](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fweight-initialization.webp&w=3840&q=75)
Weight Initialization
An essential process in neural network training that involves setting the initial values of the model's weights to influence learning effectiveness and convergence.
Generality: 675
![Prediction Error](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fprediction-error.webp&w=3840&q=75)
Prediction Error
The discrepancy between predicted outcomes by an AI model and the actual observed results in a dataset.
Generality: 675
![Node](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fnode.webp&w=3840&q=75)
Node
A fundamental unit within a neural network or graph that processes inputs to produce outputs, often reflecting the biological concept of neurons.
Generality: 500
![Batch](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fbatch.webp&w=3840&q=75)
Batch
A collection of data samples processed simultaneously in a single step of a neural network's training process.
Generality: 500
![Weight Decay](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fweight-decay.webp&w=3840&q=75)
Weight Decay
Regularization technique used in training neural networks to prevent overfitting by penalizing large weights.
Generality: 730
![Autoencoder](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fautoencoder.webp&w=3840&q=75)
Autoencoder
Type of artificial neural network used to learn efficient codings of unlabeled data, typically for the purpose of dimensionality reduction or feature learning.
Generality: 815
![Max Pooling](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fmax-pooling.webp&w=3840&q=75)
Max Pooling
Downsampling technique that reduces the dimensionality of input data by selecting the maximum value from a specified subset of the data.
Generality: 695
![Speech-to-Speech Model](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fspeech-to-speech-model.webp&w=3840&q=75)
Speech-to-Speech Model
Systems that directly convert spoken language into another language through AI, enabling real-time translation and cross-lingual communication.
Generality: 809
![SotA (State of the Art)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fsota-state-of-the-art.webp&w=3840&q=75)
SotA
State of the Art
State of the Art
The highest level of performance achieved in a specific field, particularly in AI, where it denotes the most advanced model or algorithm.
Generality: 720
![Incremental Learning](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fincremental-learning.webp&w=3840&q=75)
Incremental Learning
A method where AI systems continuously acquire new data and knowledge while retaining previously learned information without retraining from scratch.
Generality: 750
![MoE (Mixture of Experts)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fmoe-mixture-of-experts.webp&w=3840&q=75)
MoE
Mixture of Experts
Mixture of Experts
ML architecture that utilizes multiple specialist models (experts) to handle different parts of the input space, coordinated by a gating mechanism that decides which expert to use for each input.
Generality: 705
![Meta-Learning](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fmeta-learning.webp&w=3840&q=75)
Meta-Learning
Learning to learn involves techniques that enable AI models to learn how to adapt quickly to new tasks with minimal data.
Generality: 858
![Catastrophic Forgetting](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fcatastrophic-forgetting.webp&w=3840&q=75)
Catastrophic Forgetting
Phenomenon where a neural network forgets previously learned information upon learning new data.
Generality: 686
![MTL (Multi-Task Learning)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fmtl-multi-task-learning.webp&w=3840&q=75)
MTL
Multi-Task Learning
Multi-Task Learning
ML approach where a single model is trained simultaneously on multiple related tasks, leveraging commonalities and differences across tasks to improve generalization.
Generality: 761
![Wake Sleep](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fwake-sleep.webp&w=3840&q=75)
Wake Sleep
Biologically inspired algorithm used within unsupervised learning to train deep belief networks.
Generality: 540
![Transfer Learning](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Ftransfer-learning.webp&w=3840&q=75)
Transfer Learning
ML method where a model developed for a task is reused as the starting point for a model on a second task, leveraging the knowledge gained from the first task to improve performance on the second.
Generality: 870
![Continuous Learning](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fcontinuous-learning.webp&w=3840&q=75)
Continuous Learning
Systems and models that learn incrementally from a stream of data, updating their knowledge without forgetting previous information.
Generality: 870
![Early Stopping](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fearly-stopping.webp&w=3840&q=75)
Early Stopping
A regularization technique used to prevent overfitting in ML models by halting training when performance on a validation set begins to degrade.
Generality: 675
![Transfer Capability](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Ftransfer-capability.webp&w=3840&q=75)
Transfer Capability
A feature of AI systems that allows acquired knowledge in one domain or task to be applied to another distinct but related domain or task.
Generality: 775
![ReLU (Rectified Linear Unit)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Frelu-rectified-linear-unit.webp&w=3840&q=75)
ReLU
Rectified Linear Unit
Rectified Linear Unit
Activation function commonly used in neural networks which outputs the input directly if it is positive, otherwise, it outputs zero.
Generality: 855
![Classifier](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fclassifier.webp&w=3840&q=75)
Classifier
ML model that categorizes data into predefined classes.
Generality: 861
![CD (Contrastive Divergence)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fcd-contrastive-divergence.webp&w=3840&q=75)
CD
Contrastive Divergence
Contrastive Divergence
Algorithm used to approximate the gradient of the log-likelihood for training probabilistic models.
Generality: 660
![Word Vector](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fword-vector.webp&w=3840&q=75)
Word Vector
Numerical representations of words that capture their meanings, relationships, and context within a language.
Generality: 690
![Narrow AI](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fnarrow-ai.webp&w=3840&q=75)
Narrow AI
Also known as Weak AI, refers to AI systems designed to perform a specific task or a narrow range of tasks with a high level of proficiency.
Generality: 760
![DBN (Deep Belief Network)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fdbn-deep-belief-network.webp&w=3840&q=75)
DBN
Deep Belief Network
Deep Belief Network
A type of artificial neural network that is deeply structured with multiple layers of latent variables, or hidden units.
Generality: 851
![One-Shot Learning](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fone-shot-learning.webp&w=3840&q=75)
One-Shot Learning
ML technique where a model learns information about object categories from a single training example.
Generality: 542
![Model Compression](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fmodel-compression.webp&w=3840&q=75)
Model Compression
Techniques designed to reduce the size of a machine learning model without significantly sacrificing its accuracy.
Generality: 715
![Feature Learning](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Ffeature-learning.webp&w=3840&q=75)
Feature Learning
Automatically learning representations or features from raw input data in order to improve model performance and reduce dependency on manual feature engineering.
Generality: 500
![Denoising Autoencoder](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fdenoising-autoencoder.webp&w=3840&q=75)
Denoising Autoencoder
A neural network designed to reconstruct a clean input from a corrupted version, enhancing feature extraction by learning robust data representations.
Generality: 806
![Sparse Autoencoder](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fsparse-autoencoder.webp&w=3840&q=75)
Sparse Autoencoder
Type of neural network designed to learn efficient data representations by enforcing sparsity on the hidden layer activations.
Generality: 625
![Initialization](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Finitialization.webp&w=3840&q=75)
Initialization
Process of setting the initial values of the parameters (weights and biases) of a model before training begins.
Generality: 865
![Cognitive Computing](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fcognitive-computing.webp&w=3840&q=75)
Cognitive Computing
Computer systems that simulate human thought processes to solve complex problems.
Generality: 900
![AlexNet](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Falexnet.webp&w=3840&q=75)
AlexNet
Deep convolutional neural network that significantly advanced the field of computer vision by winning the ImageNet Large Scale Visual Recognition Challenge in 2012.
Generality: 610
![Dropout](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fdropout.webp&w=3840&q=75)
Dropout
Regularization technique used in neural networks to prevent overfitting by randomly omitting a subset of neurons during training.
Generality: 808
![Data Augmentation](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fdata-augmentation.webp&w=3840&q=75)
Data Augmentation
Techniques used to increase the size and improve the quality of training datasets for machine learning models without collecting new data.
Generality: 830
![Data Efficient Learning](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fdata-efficient-learning.webp&w=3840&q=75)
Data Efficient Learning
ML approach that requires fewer data to train a functional model.
Generality: 791
![Similarity Learning](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fsimilarity-learning.webp&w=3840&q=75)
Similarity Learning
A technique in AI focusing on training models to measure task-related similarity between data points.
Generality: 675
![Landmarks](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Flandmarks.webp&w=3840&q=75)
Landmarks
Key points in an image used as reference for computer vision and AI systems to understand and manipulate visual data.
Generality: 500
![Pretrained Model](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fpretrained-model.webp&w=3840&q=75)
Pretrained Model
ML model that has been previously trained on a large dataset and can be fine-tuned or used as is for similar tasks or applications.
Generality: 860
![Latent Space](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Flatent-space.webp&w=3840&q=75)
Latent Space
Abstract, multi-dimensional representation of data where similar items are mapped close together, commonly used in ML and AI models.
Generality: 805
![Embedding Space](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fembedding-space.webp&w=3840&q=75)
Embedding Space
Mathematical representation where high-dimensional vectors of data points, such as text, images, or other complex data types, are transformed into a lower-dimensional space that captures their essential properties.
Generality: 700
![Embedding](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fembedding.webp&w=3840&q=75)
Embedding
Representations of items, like words, sentences, or objects, in a continuous vector space, facilitating their quantitative comparison and manipulation by AI models.
Generality: 865
![Gradient Clipping](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fgradient-clipping.webp&w=3840&q=75)
Gradient Clipping
A technique used to mitigate the exploding gradient problem during the training of neural networks by capping gradients to a specified value range.
Generality: 625
![Model Layer](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fmodel-layer.webp&w=3840&q=75)
Model Layer
Discrete level in a neural network where specific computations or transformations are applied to the input data, progressively abstracting and refining the information as it moves through the network.
Generality: 805
![End-to-End Learning](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fend-to-end-learning.webp&w=3840&q=75)
End-to-End Learning
ML approach where a system is trained to directly map input data to the desired output, minimizing the need for manual feature engineering.
Generality: 800
![Image-to-Text Model](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fimage-to-text-model.webp&w=3840&q=75)
Image-to-Text Model
AI systems that convert visual information from images into descriptive textual representations, enabling machines to understand and communicate the content of images.
Generality: 755
![Sequence Model](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fsequence-model.webp&w=3840&q=75)
Sequence Model
Model designed to process and predict sequences of data, such as time series, text, or biological sequences.
Generality: 830
![Sequence Prediction](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fsequence-prediction.webp&w=3840&q=75)
Sequence Prediction
Involves forecasting the next item(s) in a sequence based on the observed pattern of prior sequences.
Generality: 825
![Generative AI](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fgenerative-ai.webp&w=3840&q=75)
Generative AI
Subset of AI technologies that can generate new content, ranging from text and images to music and code, based on learned patterns and data.
Generality: 830
![Discriminative AI](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fdiscriminative-ai.webp&w=3840&q=75)
Discriminative AI
Algorithms that learn the boundary between classes of data, focusing on distinguishing between different outputs given an input.
Generality: 840
![Recognition Model](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Frecognition-model.webp&w=3840&q=75)
Recognition Model
Element of AI that identifies patterns and features in data through learning processes.
Generality: 790
![Autoregressive Sequence Generator](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fautoregressive-sequence-generator.webp&w=3840&q=75)
Autoregressive Sequence Generator
A predictive model harnessed in AI tasks, particularly involving times series, which leverages its own prior outputs as inputs in subsequent predictions.
Generality: 650
![Sequential Models](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fsequential-models.webp&w=3840&q=75)
Sequential Models
Type of data models in AI where the arrangement of data points or events adhere to a specific order for predictive analysis and pattern recognition.
Generality: 815
![Convergence](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fconvergence.webp&w=3840&q=75)
Convergence
The point at which an algorithm or learning process stabilizes, reaching a state where further iterations or data input do not significantly alter its outcome.
Generality: 845
![Overparameterized](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Foverparameterized.webp&w=3840&q=75)
Overparameterized
ML model that has more parameters than the number of data points available for training.
Generality: 555
![Loss Landscape](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Floss-landscape.webp&w=3840&q=75)
Loss Landscape
The topographical representation of a neural network's loss function showcasing the variations in loss values across different parameter settings.
Generality: 500
![Model Distillation](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fmodel-distillation.webp&w=3840&q=75)
Model Distillation
ML technique where a larger, more complex model (teacher) is used to train a smaller, simpler model (student) to approximate the teacher's predictions while maintaining similar performance.
Generality: 625
![Teacher Model](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fteacher-model.webp&w=3840&q=75)
Teacher Model
Pre-trained, high-performing model that guides the training of a simpler, student model, often in the context of knowledge distillation.
Generality: 561
![Activation Data](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Factivation-data.webp&w=3840&q=75)
Activation Data
Intermediate outputs produced by neurons in a neural network when processing input data, which are used to evaluate and update the network during training.
Generality: 575
![DQN (Deep Q-Networks)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fdqn-deep-q-networks.webp&w=3840&q=75)
DQN
Deep Q-Networks
Deep Q-Networks
RL technique that combines Q-learning with deep neural networks to enable agents to learn how to make optimal decisions from high-dimensional sensory inputs.
Generality: 853
![Variance Scaling](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fvariance-scaling.webp&w=3840&q=75)
Variance Scaling
Variance scaling is a technique used in machine learning to ensure weights of layers are initialized in a way that maintains consistent variance of activations throughout a neural network.
Generality: 525
![GLU (Gated Linear Unit)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fglu-gated-linear-unit.webp&w=3840&q=75)
GLU
Gated Linear Unit
Gated Linear Unit
Neural network component that uses a gating mechanism to control information flow, improving model efficiency and performance.
Generality: 665
![Layer Normalization](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Flayer-normalization.webp&w=3840&q=75)
Layer Normalization
Technique used in neural networks to normalize the inputs across the features within a layer, improving training stability and model performance, particularly in recurrent and transformer models.
Generality: 715
![Black Box Problem](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fblack-box-problem.webp&w=3840&q=75)
Black Box Problem
The difficulty in understanding and interpreting how an AI system, particularly ML models, makes decisions.
Generality: 850
![Few Shot](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Ffew-shot.webp&w=3840&q=75)
Few Shot
ML technique designed to recognize patterns and make predictions based on a very limited amount of training data.
Generality: 675
![Federated Learning](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Ffederated-learning.webp&w=3840&q=75)
Federated Learning
ML approach enabling models to be trained across multiple decentralized devices or servers holding local data samples, without exchanging them.
Generality: 805
![Inference Acceleration](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Finference-acceleration.webp&w=3840&q=75)
Inference Acceleration
Methods and hardware optimizations employed to increase the speed and efficiency of the inference process in machine learning models, particularly neural networks.
Generality: 775
![Multimodal](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fmultimodal.webp&w=3840&q=75)
Multimodal
AI systems or models that can process and understand information from multiple modalities, such as text, images, and sound.
Generality: 837
![Responsible AI](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fresponsible-ai.webp&w=3840&q=75)
Responsible AI
Application of AI in a manner that is transparent, unbiased, and respects user privacy and value.
Generality: 815
![Federated Training](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Ffederated-training.webp&w=3840&q=75)
Federated Training
Decentralized machine learning approach where multiple devices or nodes collaboratively train a shared model while keeping their data localized, rather than aggregating it centrally.
Generality: 805
![Capsule Networks](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fcapsule-networks.webp&w=3840&q=75)
Capsule Networks
Type of artificial neural network designed to improve the processing of spatial hierarchical information by encoding data into small groups of neurons called capsules.
Generality: 660
![Expressive Hidden States](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fexpressive-hidden-states.webp&w=3840&q=75)
Expressive Hidden States
internal representations within a neural network that effectively capture and encode complex patterns and dependencies in the input data.
Generality: 695
![Neurosymbolic AI](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fneurosymbolic-ai.webp&w=3840&q=75)
Neurosymbolic AI
Integration of neural networks with symbolic AI to create systems that can both understand and manipulate symbols in a manner similar to human cognitive processes.
Generality: 675
![Out of Distribution](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fout-of-distribution.webp&w=3840&q=75)
Out of Distribution
Data that differs significantly from the training data used to train a machine learning model, leading to unreliable or inaccurate predictions.
Generality: 675
![Point-wise Feedforward Network](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fpoint-wise-feedforward-network.webp&w=3840&q=75)
Point-wise Feedforward Network
Neural network layer that applies a series of linear and non-linear transformations to each position (or
Generality: 625
![Hybrid AI](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fhybrid-ai.webp&w=3840&q=75)
Hybrid AI
Combines symbolic AI (rule-based systems) and sub-symbolic AI (machine learning) approaches to leverage the strengths of both for more versatile and explainable AI systems.
Generality: 820
![Masking](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fmasking.webp&w=3840&q=75)
Masking
Technique used in NLP models to prevent future input tokens from influencing the prediction of current tokens.
Generality: 639
![Ablation](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fablation.webp&w=3840&q=75)
Ablation
Method where components of a neural network are systematically removed or altered to study their impact on the model's performance.
Generality: 650
![Zero-shot Capability](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fzero-shot-capability.webp&w=3840&q=75)
Zero-shot Capability
The ability of AI models to perform tasks or make predictions on new types of data that they have not encountered during training, without needing any example-specific fine-tuning.
Generality: 775
![SSL (Self-Supervised Learning)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fssl-self-supervised-learning.webp&w=3840&q=75)
SSL
Self-Supervised Learning
Self-Supervised Learning
Type of ML where the system learns to predict part of its input from other parts, using its own data structure as supervision.
Generality: 815
![LLM (Large Language Model)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fllm-large-language-model.webp&w=3840&q=75)
LLM
Large Language Model
Large Language Model
Advanced AI systems trained on extensive datasets to understand, generate, and interpret human language.
Generality: 827
![Next Token Prediction](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fnext-token-prediction.webp&w=3840&q=75)
Next Token Prediction
Technique used in language modeling where the model predicts the following token based on the previous ones.
Generality: 735
![xLSTM](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fxlstm.webp&w=3840&q=75)
xLSTM
Extended form of Long Short-Term Memory (LSTM), integrating enhancements for scalability and efficiency in DL models.
Generality: 675
![Base Model](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fbase-model.webp&w=3840&q=75)
Base Model
Pre-trained AI model that serves as a starting point for further training or adaptation on specific tasks or datasets.
Generality: 790
![DLMs (Deep Language Models)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fdlms-deep-language-models.webp&w=3840&q=75)
DLMs
Deep Language Models
Deep Language Models
Advanced ML models designed to understand, generate, and translate human language by leveraging DL techniques.
Generality: 874
![Self-Supervised Pretraining](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fself-supervised-pretraining.webp&w=3840&q=75)
Self-Supervised Pretraining
ML approach where a model learns to predict parts of the input data from other parts without requiring labeled data, which is then fine-tuned on downstream tasks.
Generality: 725
![Adapter Layer](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fadapter-layer.webp&w=3840&q=75)
Adapter Layer
Neural network layer used to enable transfer learning by adding small, trainable modules to a pre-trained model, allowing it to adapt to new tasks with minimal additional training.
Generality: 625
![Continual Pre-Training](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fcontinual-pre-training.webp&w=3840&q=75)
Continual Pre-Training
Process of incrementally training a pre-trained ML model on new data or tasks to update its knowledge without forgetting previously learned information.
Generality: 670
![Post-Training](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fpost-training.webp&w=3840&q=75)
Post-Training
Techniques and adjustments applied to neural networks after their initial training phase to enhance performance, efficiency, or adaptability to new data or tasks.
Generality: 650
![AMI (Advanced Machine Intelligence)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fami-advanced-machine-intelligence.webp&w=3840&q=75)
AMI
Advanced Machine Intelligence
Advanced Machine Intelligence
Refers to high-level AI systems possessing the capability to perform complex cognitive tasks with or without human-like reasoning.
Generality: 873
![Scaling Laws](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fscaling-laws.webp&w=3840&q=75)
Scaling Laws
Mathematical relationships that describe how the performance of machine learning models, particularly deep learning models, improves as their size, the amount of data, or computational resources increases.
Generality: 835
![1-N Systems](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2F1-n-systems.webp&w=3840&q=75)
1-N Systems
Architectures where one input or controller manages multiple outputs or agents, applicable in fields like neural networks and robotics.
Generality: 790
![Scaling Hypothesis](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fscaling-hypothesis.webp&w=3840&q=75)
Scaling Hypothesis
Enlarging model size, data, and computational resources can consistently improve task performance up to very large scales.
Generality: 765
![Generative Model](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fgenerative-model.webp&w=3840&q=75)
Generative Model
A type of AI model that learns to generate new data instances that mimic the training data distribution.
Generality: 840
![Parametric Knowledge](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fparametric-knowledge.webp&w=3840&q=75)
Parametric Knowledge
Information and patterns encoded within the parameters of a machine learning model, which are learned during the training process.
Generality: 849
![Instruction Following Model](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Finstruction-following-model.webp&w=3840&q=75)
Instruction Following Model
AI system designed to execute tasks based on specific commands or instructions provided by users.
Generality: 640
![Instruction-Following](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Finstruction-following.webp&w=3840&q=75)
Instruction-Following
Ability to accurately understand and execute tasks based on given directives.
Generality: 725
![Transformative AI](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Ftransformative-ai.webp&w=3840&q=75)
Transformative AI
AI systems capable of bringing about profound, large-scale changes in society, potentially altering the economy, governance, and even human life itself.
Generality: 825
![Self-Correction](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fself-correction.webp&w=3840&q=75)
Self-Correction
An AI system's ability to recognize and rectify its own mistakes or errors without external intervention.
Generality: 815
![VLM (Visual Language Model)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fvlm-visual-language-model.webp&w=3840&q=75)
VLM
Visual Language Model
Visual Language Model
AI models designed to interpret and generate content by integrating visual and textual information, enabling them to perform tasks like image captioning, visual question answering, and more.
Generality: 621
![Foundation Model](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Ffoundation-model.webp&w=3840&q=75)
Foundation Model
Type of large-scale pre-trained model that can be adapted to a wide range of tasks without needing to be trained from scratch each time.
Generality: 835
![MLLMs (Multimodal Large Language Models)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Fmllms-multimodal-large-language-models.webp&w=3840&q=75)
MLLMs
Multimodal Large Language Models
Multimodal Large Language Models
Advanced AI systems capable of understanding and generating information across different forms of data, such as text, images, and audio.
Generality: 625
![LVLMs (Large Vision Language Models)](/vocab/_next/image?url=%2Fvocab%2Fimages%2Farticles%2Fsmall%2Flvlms-large-vision-language-models.webp&w=3840&q=75)
LVLMs
Large Vision Language Models
Large Vision Language Models
Advanced AI systems designed to integrate and interpret both visual and textual data, enabling more sophisticated understanding and generation based on both modalities.
Generality: 675