glossary

Key terms and definitions


Agent
An AI system that uses an observe-think-act loop with tool calling to autonomously accomplish tasks.
Attention
A mechanism that allows neural networks to focus on relevant parts of the input when producing output.
BERT
Bidirectional Encoder Representations from Transformers — an encoder-only model pre-trained on masked language modelling and next sentence prediction.
Embedding
A learned dense vector representation that maps discrete tokens to continuous vector spaces.
GPT
Generative Pre-trained Transformer — a family of decoder-only language models that generate text by predicting the next token.
GRU
Gated Recurrent Unit — a simplified RNN variant with fewer gates than LSTM that achieves comparable performance with less computation.
JSON Schema
A vocabulary for annotating and validating JSON documents, used to define tool interfaces for language models.
LSTM
Long Short-Term Memory — an RNN variant with gating mechanisms that can learn long-range dependencies in sequential data.
NLP
Natural Language Processing — the field of AI concerned with enabling computers to understand, interpret, and generate human language.
Prompt Injection
An adversarial attack where malicious input manipulates a language model into performing unintended actions.
RNN
Recurrent Neural Network — a neural network architecture that processes sequences by maintaining a hidden state across time steps.
Tool Calling
A mechanism that allows large language models to invoke external functions and APIs by generating structured requests.
Transformer
A neural network architecture based on self-attention mechanisms, introduced in 'Attention Is All You Need' (2017).