LLM Terms Flashcards

1
Q

Large Language Model (LLM)

A

A type of neural network model designed to understand and generate human-like text by learning from vast amounts of language data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Transformer Architecture

A

A deep learning model architecture that uses self-attention mechanisms to process input data in parallel, making it the foundation of modern LLMs like GPT and BERT.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Self-Attention

A

A mechanism that enables the model to focus on different parts of the input sequence, assigning importance scores to each token relative to others in the context.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Context Length

A

The maximum number of tokens an LLM can process in a single input. Longer context lengths allow the model to capture more extensive dependencies in the input text.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Tokenization

A

The process of converting text into smaller units (tokens) that the LLM can process, such as words, subwords, or characters.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Subword Tokenization

A

A method of breaking words into smaller units (subwords), enabling the model to handle out-of-vocabulary words and morphological variations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Byte-Pair Encoding (BPE)

A

A tokenization technique that iteratively merges the most frequent pairs of characters or character sequences to create subword tokens.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Prompt Engineering

A

The practice of crafting input prompts to guide the LLM in generating specific, desired outputs, essential for steering model behavior in applications.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Few-Shot Learning

A

A model’s ability to learn and generalize from only a few examples provided in the prompt, enabling the generation of contextually relevant responses.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Zero-Shot Learning

A

The capability of an LLM to perform a task without explicit examples in the prompt, relying instead on its pre-trained knowledge.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Fine-Tuning

A

The process of training a pre-trained LLM on a smaller, task-specific dataset to adapt the model for specific applications.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Retrieval-Augmented Generation (RAG)

A

A technique combining information retrieval and LLMs, where relevant documents are retrieved and provided as context to the model during inference.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Knowledge Base

A

A structured repository of information used to provide factual data for an LLM, enhancing its ability to generate accurate and contextually appropriate responses.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Semantic Search

A

A search method that leverages the meaning of words and their relationships to find relevant documents or information, often used in conjunction with LLMs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Embeddings

A

Dense vector representations of words, sentences, or documents that capture their semantic meanings and are used for tasks like semantic search in LLM applications.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Memory Management

A

Techniques used in LLMs to handle the model’s context length, including how to store and retrieve information across long conversations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Temperature

A

A parameter that controls the randomness of an LLM’s output. Higher values generate more diverse outputs, while lower values make responses more deterministic.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Top-k Sampling

A

A decoding method that restricts the model’s predictions to the top k most likely tokens at each step, enhancing the quality of generated text.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Top-p (Nucleus) Sampling

A

A sampling method that dynamically selects tokens based on their cumulative probability distribution, allowing for more flexible and coherent text generation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Beam Search

A

A decoding strategy that explores multiple possible sequences during text generation, keeping the top n most probable sequences at each step.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Hallucination

A

A phenomenon where an LLM generates content that is not factually accurate or related to the given context, often seen as a critical challenge in LLM applications.

22
Q

Latent Space

A

The high-dimensional space in which LLMs represent input data, capturing complex relationships and semantics between words and sentences.

23
Q

Embedding Models

A

Models trained to generate vector representations of input data (words, sentences, documents), often used in combination with LLMs for tasks like semantic search.

24
Q

Sequence-to-Sequence (Seq2Seq) Model

A

A model architecture that maps input sequences to output sequences, commonly used in tasks like translation and summarization.

25
Q

Attention Mask

A

A mechanism that controls which parts of the input sequence an LLM should focus on, used to handle padding and manage context in sequence processing.

26
Q

Positional Encoding

A

A technique used in Transformer models to provide information about the order of tokens in the input sequence.

27
Q

Encoder-Decoder Architecture

A

A neural network framework that consists of an encoder processing the input sequence and a decoder generating the output sequence, foundational to many NLP tasks.

28
Q

Pre-training

A

The initial phase of training an LLM on a large corpus of text data to learn general language patterns and representations.

29
Q

Transfer Learning

A

Adapting a pre-trained LLM to a new task or domain by fine-tuning it on a smaller, specialized dataset.

30
Q

Language Model Inference

A

The process of generating output text from an LLM given an input prompt or query.

31
Q

Hyperparameter Tuning

A

Adjusting model parameters (e.g., learning rate, batch size, temperature) to optimize the performance of LLMs for specific tasks.

32
Q

Model Checkpointing

A

Saving intermediate states of an LLM during training, allowing for recovery and further fine-tuning if necessary.

33
Q

Gradient Accumulation

A

A technique used to overcome memory limitations during LLM training by accumulating gradients over multiple batches before updating model weights.

34
Q

Distributed Training

A

The process of training an LLM across multiple GPUs or machines to handle large datasets and model sizes efficiently.

35
Q

Data Augmentation

A

Techniques for increasing the diversity of training data, such as paraphrasing or adding noise, to improve LLM performance.

36
Q

Text Summarization

A

The task of generating a concise and coherent summary of a longer text, often facilitated by LLMs in various applications.

37
Q

Named Entity Recognition (NER)

A

A technique for identifying and classifying entities (e.g., names, dates, locations) in text, used to extract structured information in LLM applications.

38
Q

Conversational Agent

A

A system that uses an LLM to interact with users in natural language, often designed to handle tasks like customer support, information retrieval, or personal assistance.

39
Q

Conversational Memory

A

Mechanisms in an LLM application that retain context across multiple user interactions to facilitate coherent dialogue.

40
Q

Latency

A

The time delay between a user input and the model’s response, a critical factor in the user experience of LLM-based applications.

41
Q

Inference Optimization

A

Techniques such as model quantization or pruning to reduce the computational cost and latency of LLMs during inference.

42
Q

Knowledge Distillation

A

The process of transferring knowledge from a large, complex LLM to a smaller, more efficient model to improve performance in resource-constrained environments.

43
Q

Prompt Tuning

A

The process of adjusting prompts used with an LLM to elicit more accurate, relevant, or contextually appropriate responses.

44
Q

Bias in Language Models

A

The presence of stereotypes or unfair associations learned from training data, necessitating strategies to mitigate bias in LLM applications.

45
Q

Zero-Shot Transfer

A

The ability of an LLM to perform tasks in new domains without explicit retraining on domain-specific data.

46
Q

Conversational Turn

A

A single interaction between a user and an LLM in a dialogue, which includes both the user input and the model’s response.

47
Q

Contextual Embeddings

A

Word embeddings that change depending on the context in which a word appears, enabling more nuanced understanding in LLMs.

48
Q

Query Parsing

A

The process of analyzing and interpreting a user’s input to understand intent and extract relevant information for the LLM application.

49
Q

Memory Augmented LLM

A

An LLM equipped with external memory components to store and retrieve information beyond its fixed parameters, enhancing its capability to answer specific queries.

50
Q

Scalability

A

The ability of an LLM application to handle an increasing amount of work, data, or users without sacrificing performance or accuracy.