Midterm Flashcards

1
Q

SOTA

A

State of the Art

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Clark-Fisher Hypothesis

A

As jobs get displaced people will move sectors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Turing Test

A

Test to see how well a robot can imitate a juman

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

AI winter

A

period of time where people weren’t investing or researching AI

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Expert Systems

A

Machine expert in a topic, usually consists of a knowledge base and an inference engine

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

knowledge engineer

A

responsible for designing and maintaining expert systems, they gather knowledge from human experts to turn into rules or facts from the system

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

knowledge base

A

What the AI system knows. Ex: rules, facts, information about the role of the AI, etc.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Polanyi principle

A

We often know more than we can articulate.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

backpropogation

A

minimizes error in predictions made by the network. The network can learn from its mistakes and adjust weights between neurons to create better inferences in the future

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Artificial neural network

A

computational model of layers of nodes or neurons

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

GPU

A

graphics processing unit

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

FLOPS

A

Floating point calculations a computer can perform in one second

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Tops

A

Trillions of floating point calculations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

RNN

A

Recurrent neural networks are great handling sequential data and finding patterns

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

sequence to sequence

A

A model that takes one sequence of data, like an english sentence, and translates it into a different sequence, like a portuguese sentence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

long short-term memory

A

a type of RNN achitecture designed to effectively learn from and remember long sequences of data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

transformer architecture

A

Through the use of a “self-attention” mechanism, it weighs the importance of words, better understanding relationships and context

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

parallel processing

A

simultaneous execution of multiple tasks or processes. Helps with handling large amounts of data at once, instead of working through it sequentially

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

emergent behavior

A

when the AI does something unexpected because of all the pieces working together

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

AGI

A

artificial general intelligence- an AI that can understand, learn, and apply knowledge in many ways

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Singularity

A

a point when AIs become more intelligent than humans, leading to rapid tech advancements and societal changes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

GenAi

A

Ai that can generate new content

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

machine learning

A

focuses on development of algorithms and statistical models that help computers learn from and make predictions or decisions based on data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

deep learning

A

specialized area within machine learning that uses neural networks with many layers to analyze and interpret complex data patterns. Mimcs the human brain in tasks like image and speech recognition and languages

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

data features

A

individual measurable properties or characteristics of the data used in MLMs. They are the inputs that algorithms analyze

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

perceptron

A

an artificial neuron and one of the simplest forms of a neural network

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

activation function

A

mathematical function used in neural networks to determine the output of a neuron based on the input

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

Parameters

A

internal variables of a model that are learned from training

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

multilayer perceptron

A

(MLP) a type of artificial neural network that consists of multiple layers of nodes, or neurons, including an input layer, one or more hidden layers, and an output layer

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

feature detection

A

process of identifying and extracting important characteristics or patterns from raw data, which can be used for further analysis or modeling

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

ANN model loss

A

a measure of how well an ANN is performing during training. It quantifies the difference between predicted outputs and actual target values from the training data. Lower loss is better

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

Hallucinations

A

when AI models generate fake made up outcomes that sound plausible

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

Self supervised learning

A

machine learning where models predict parts of the data from other parts, without needing labeled data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

Training rate

A

a hyperparameter in machine learning that determines how much model’s weights are updated during training in response to the calculated error.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

LLM completion

A

the process of generating text based on a prompt

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
36
Q

open source

A

publicly available for modification, use, and distribution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
37
Q

Fine-tuning for instruction

A

Take a pre-trained mode, and add additional training on a specific topic/dataset

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
38
Q

Model alignment

A

ensuring AI outputs correspond with human values and expectations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
39
Q

Supervised learning

A

model is trained on labeled data. Clear examples of desired outcomes helps the model make accurate predictions in real-world scenarios

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
40
Q

Reinforcement Learning from human Feedback

A

machine learning with human feedback instead of solely predefined rewards

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
41
Q

Foundation model

A

large scale model trained on big data, commonly used as parent models

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
42
Q

Semantic embedding

A

a technique that transforms words, phrases, or documents into vectors in a high-dimensional space, capturing meanings and relationships

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
43
Q

attention mechanism

A

a technique where models focus on specific parts of input data when making predictions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
44
Q

Autoregression

A

the model predicts the next value in a sequence based on the previous values

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
45
Q

digital twin

A

a virtual representation of a physical object, system, or process that uses real-time data to simulate its behavior and performance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
46
Q

Latency

A

time delay between the input being provided to an AI system and the output being generated (includes processing time, poor connection, etc)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
47
Q

AI use case

A

specific application or scenario where ai tech is implemented to solve a particular problem

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
48
Q

GLUE benchmark

A

general language understanding evaluation; assess models on a variety of language understanding tasks; language comprehension skills

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
49
Q

SQuAD Benchmark

A

stanford question answering dataset; reading comprehension and question answering; given passages can it extract certain answers; retrieval skills

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
50
Q

RACE benchmark

A

reading comprehension from examinations. evaluate reading comprehension, can AI mimic human-like understanding of complex texts; understand and interpret text

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
51
Q

5 ai accuracy metrics

A

quantititative metrics;
1. Accuracy 2. Precision
3. Recall (sensitivity)
4. F1 Score
5. AUC-ROC

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
52
Q

AI model error rate

A

frequency of incorrect predictions made by an AI model; lower error rate is more accurate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
53
Q

Model Perplexity

A

how well a probability model predicts a sample; quantifies uncertainty or unpredictability of a model when generating text; lower perplexity score= model is better at predicting the next word in a sequence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
54
Q

BLEU score

A

bilingual evaluation understudy; machine translation. 0-1 with 1 being better

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
55
Q

AI Recall/sensitivity

A

how many of the actual positive cases the model successfully identified; important where missing a positive (like a medical diagnosis) is really important

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
56
Q

F1 score

A

combines precision and recall; how well a model performs when both false positives and false negatives are important; high F1 score is good; precision (accuracy of positive precisions); recall (ability to identify all positive cases)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
57
Q

AI model Efficiency

A
  1. Inference time: make predictions time
  2. Model size: amount of memory required to store the model
  3. energy consumption: how much energy the model uses during training and inference
  4. throughput: number of predictions the model can make in a given time period
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
58
Q

key aspects of load handling

A

ability of an AI system to manage and process varying amounts of data or requests efficiently;
1. scalability
2. throughput
3. latency
4. Load balancing: distributing incoming requests across multiple instances
5. fault tolerance: continue functioning correctly even when some components fail

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
59
Q

Kubernetes

A

automates deployment, scaling, and management of containerized applications; manages resources

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
60
Q

Deployment elasticity

A

ability to dynamically adjust its resources based on demand; efficiently handle varying workloads

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
61
Q

Needle in a haystack

A

the challenge of finding a specific piece of valuable information or insight within a vast amount of data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
62
Q

Agentic workflow

A

AI agents autonomously perform tasks and make decisions with a defined workflow, work with human users too

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
63
Q

Low rank adapters

A

instead of retraining an entire model, low-rank adapters introduce small, trainable modules or adapters that can be inserted into existing architecture

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
64
Q

context size/window

A

amount of info/tokens a model can consider at one time when processing input data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
65
Q

FLOPS and TOPS

A

FLOPS: floating point operations per second; measure of a computer’s performance on complex calculations: how well can we train and deploy these ai systems

TOPS: trillions of operations per second; measure performance of computing systems; comparing AI hardware

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
66
Q

Model quantization

A

reduce the size and computational requirements of models; more efficient

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
66
Q

Edge computing

A

process data closer to the source of data generation rather than centralized data centers; allows for quicker data processing and less latency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
67
Q

LLM distillation

A

create smaller more efficient versions of LLMs while retaining much of the performance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
68
Q

Hugging face

A

Open-source tools and models; large transformers library

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
69
Q

Interpreted software code

A

programming code that is executed line-by-line by an interpreter rather than being compiled into machine code beforehand; easier to read, but slower than compiled code

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
70
Q

compiled software code

A

programming code that is transformed into machine code by a compiler before it’s executed; creates a file that a computer can run directly on its hardware

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
71
Q

Assembly language

A

low-level programming language that is close to machine code (similar to binary); harder to write in but more customizable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
72
Q

binary executable

A

machine readable file that doesn’t need further translation or interpretation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
73
Q

MIT License

A

permissive free software license that allows developers to use, modify, and distribute software with few restrictions; just need to include the original copyright notice

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
74
Q

Apache 2.0 license

A

open source software license; no need to disclose their own source code; proper attribution required; more terms than MIT, doesn’t require derivative works to be open source

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
75
Q

GNU general public license (GPL)

A

copyleft license where any modified versions of the software must also be distributed under the GPL license

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
76
Q

GNU Lesser General Public License (LGPL)

A

allows LGPL software to be linked with proprietary software without forcing the proprietary software to be also licensed LGPL; can take advantage of LGPT software without giving up proprietary rights

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
77
Q

OpenRAIL

A

free use and redistribution of software, but restricted to ethical uses

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
78
Q

LMOps

A

analogous to DevOps in software development. processes involved in managing, deploying, and maintaining language models

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
79
Q

Prompt caching

A

store previously used prompts and their corresponding responses to retrieve those responses for future prompts

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
80
Q

attention mechanism

A

model focuses on specific parts of the input data when generating an output; the model learns to weigh the importance of different parts of the input based on the context of the task; weighted sums

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
81
Q

positional encoding

A

indicates position/order of elements in a sequence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
82
Q

JSON object

A

javascript; easy for us to read and easy for machines to generate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
83
Q

AI temperature

A

how random the responses are

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
84
Q

Run AI Model Locally

A

Executing an AI algorithm directly on a local machine (e.g., a laptop or on-prem server), without sending data to the cloud. This allows for faster inference times, improved data privacy, and offline capabilities but often requires more powerful hardware.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
85
Q

Host computer

A

the primary machine that runs virtual machines, containers, or locally hosted AI models

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
86
Q

Edge computing

A

where data is processed at or near the source of generation, rather than in centralized cloud servers. essential for low latency applications

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
87
Q

User interface

A

the point of interaction between the user and a computer system; chatbots, dashboards, apps

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
88
Q

Client server configuration

A

a network where a client makes requests to a centralized server. common in web based ai tools where the model is on the server and accessed via the web

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
89
Q

Front end software

A

the visible and interactive part of a software application that users interact with. handles the presentation layer and sends inputs to the back-end

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
90
Q

Back end software

A

handles business logic, database interactions, and processing tasks, back end hosts models, processed input data, and returns data to the front end

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
91
Q

Hosting provider

A

a company or platform that offers infrastructure for storing and running software or websites. ex: google

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
92
Q

Software environment

A

the setup in which a software application runs, including operating system, libraries, dependencies, and configurations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
93
Q

Software as a service (SaaS)

A

a cloud-based delivery model where users access applications via a web browser without installing anything locally

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
94
Q

Server farm

A

a group of networked servers housed in one location used to provide large-scale computing resources. ai models are trained and deployed using such server clusters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
95
Q

Economies of scale

A

the cost advantages that arise with increased production; where cloud providers operate vast server farms to reduce the per unit cost of computation and storage

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
96
Q

Hypervisor

A

software that enables virtualization by allowing multiple virtual machines to run on a single host. it allocates hardware resources to each VM

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
97
Q

Virtual machines (VMs)

A

software emulations of physical computers. each VM has its own OS and can run applications independently. useful for isolating environments in ai experiments

98
Q

Virtual Private Servers

A

a virtual server sold as a service by hosting providers. offers more control and isolation than shared hosting and is often used for hosting ai models with moderate traffic

99
Q

Containerization

A

a lightweight form of virtualization that packages software and its dependencies into containers. unlike vm’s containers share the host OS, making them faster and more efficient

100
Q

Docker

A

a popular platform for building, packaging, and running containerized applications. used in ai development to ensure models are portable and consistent

101
Q

multi cloud software deployment

A

deploying applications across multiple cloud platforms. reduces reliance on a single provider and can optimize performance and cost for ai services

102
Q

Kubernetes

A

an open-source system for automating the deployment, scaling, and management of containerized applications. used in managing ai services at scale

103
Q

Kubernetes auto scaling

A

a feature that automatically adjusts the number of containers or resources based on demand.

104
Q

ai model latency

A

the time it takes for an ai model to return a result after receiving input. lower latency is crucial for real-time applications like voice assistants or fraud detection systems

105
Q

self hosting

A

deploying and managing software or ai models on your own infrastructure instead of using cloud providers. more control and privacy, but requires more technical know-how

106
Q

SERP (Search engine results page)

A

the page displayed by search engines in response to the user’s query, ai ranks results

107
Q

middleware

A

software that acts as a bridge between different systems or applications. connects the UI to the AI model= example

108
Q

Thin client

A

a minimal computing device that relies heavily on a central server for processing. often used in enterprise settings with centralized ai processing

109
Q

thick client

A

fully functional device that performs most computing tasks locally. clients don’t depend on constant server access

110
Q

microservice

A

a small, independent component of a larger application that performs one function. ai models are often deployed as microservices so they can be updated or scaled independently

111
Q

Poe AI platform

A

allows users to interact with AI models from various providers (like OpenAI, Anthropic, etc.) in a unified chat interface. It supports integration and comparison of model outputs.

112
Q

Zapier AI Platform

A

A no-code automation platform that integrates various software tools. With AI integrations, it can automate tasks like generating content or categorizing inputs using language models.

113
Q

AI system integrations

A

The process of embedding AI functionality into existing systems like CRM, ERP, or web applications. This requires APIs, middleware, and thoughtful workflow design.

114
Q

Ollama

A

A platform for running large language models locally. It simplifies downloading, running, and experimenting with AI models like LLaMA on your own machine.

115
Q

GPU (graphics processingn unit)

A

Originally designed for rendering graphics, GPUs are now heavily used for AI model training and inference due to their parallel processing capabilities.

116
Q

GPU Cluster

A

A group of GPUs working together, often distributed across machines, to speed up AI computations, especially model training and large batch inference.

117
Q

Tensor Processing Units (TPUs)

A

Custom hardware by Google designed specifically for AI and machine learning workloads. They’re optimized for matrix operations typical in neural networks.

118
Q

Neural Processing Units (NPUs)

A

Chips designed to accelerate neural network processing, commonly found in mobile and edge devices to allow local AI inference with low power consumption.

119
Q

SRS Document (software requirements specification)

A

A detailed description of a software system’s functionality, constraints, and environment. It outlines what the AI system should do, including inputs, outputs, and performance expectations—serving as a contract between stakeholders and developers.

120
Q

DevOps

A

A combination of “Development” and “Operations,” DevOps is a set of practices that aims to automate and integrate software development and IT operations. It promotes continuous integration, continuous delivery (CI/CD), collaboration between teams, and faster release cycles—critical for AI tools needing constant updates and improvements.

121
Q

Software deployment elasticity

A

The system’s ability to dynamically allocate or release resources (e.g., computing power or memory) based on current demand. In AI deployments, elasticity ensures cost-effectiveness by scaling resources up during high traffic and down during low usage.

122
Q

Software deployment scaling

A

The ability of an application to handle increased workload or user traffic by upgrading its infrastructure. This can be vertical (more powerful machines) or horizontal (more instances). AI services benefit from scaling when many users query models simultaneously.

123
Q

Software Latency

A

The delay between a user’s action (like submitting a prompt) and the software’s response. In AI systems, latency is affected by model size, compute power, network conditions, and system design. Lower latency = better user experience.

124
Q

software usability

A

Refers to how easy and efficient it is for users to interact with a software application. In AI, good usability means non-technical users can still get accurate results or insights without needing deep knowledge of machine learning.

125
Q

Prompt engineering

A

The art and science of crafting input prompts to get desired outputs from AI models—especially large language models (LLMs). Effective prompt engineering considers wording, structure, examples, and constraints

126
Q

System prompt

A

A predefined instruction embedded into the AI model to set its tone, behavior, or scope (e.g., “You are a helpful assistant”). It shapes how the model interprets and responds to user queries.

127
Q

Ai model temperature metaparameter

A

A setting that controls the randomness of an AI model’s output. A low temperature (e.g., 0.2) makes answers more focused and deterministic. A high temperature (e.g., 0.9) introduces more creativity and variation.

128
Q

AI Playground

A

An interactive web environment for testing and experimenting with AI models, often provided by companies like OpenAI or Hugging Face. These tools help users understand how AI models behave and respond to prompts.

129
Q

No-code frameworks

A

Platforms that let users build apps, workflows, or AI integrations without writing code. They’re useful for business teams and operations managers to implement automation or deploy AI tools quickly (e.g., Zapier, Bubble).

130
Q

Low code frameworks

A

Platforms that require minimal coding to build software. They offer more flexibility than no-code while still reducing the need for deep technical knowledge. Ideal for rapidly prototyping AI solutions.

131
Q

AI Model agnostic

A

Refers to software or platforms that work with multiple types of AI models, regardless of their architecture or provider. This enables flexibility to swap models like GPT-4, Claude, or LLaMA based on use case or performance.

132
Q

RAG (retrieval-augmented generation)

A

An AI technique where the model pulls information from external sources (like a database or document store) before generating a response. This improves factual accuracy and enables dynamic answers grounded in current or domain-specific data.

133
Q

Process workflows

A

A defined sequence of tasks or steps within a system or business process. AI can automate, monitor, or optimize workflows to save time, reduce errors, and increase scalability.

134
Q

MindStudio AI

A

A platform (typically no-code or low-code) designed to create and deploy custom AI workflows or apps. It allows users to build AI tools, chatbots, or agents by configuring behavior without deep programming knowledge.

135
Q

Google Colab

A

A cloud-based Jupyter notebook environment that supports Python and GPU/TPU acceleration. Popular for prototyping machine learning models and testing AI code collaboratively.

136
Q

Software Package

A

A bundled set of code, libraries, and configuration files that can be installed to add functionality.

137
Q

JSON Format

A

Stands for JavaScript Object Notation. A lightweight data-interchange format that is easy to read and write. AI systems use JSON to structure inputs/outputs, store model responses, and send data via APIs.

138
Q

Software Endpoints

A

URLs or addresses where APIs receive requests and return responses. In AI, endpoints often refer to access points for querying a model

139
Q

API Key

A

A unique string used to authenticate a user or application when accessing a software API. In AI systems, an API key ensures secure access to model endpoints and usage tracking.

140
Q

LLM Toxicity

A

Refers to the presence of harmful, offensive, or biased language in responses generated by Large Language Models (LLMs). Managing toxicity is essential for maintaining ethical AI usage

141
Q

LLM Moderation

A

The techniques and systems used to filter or flag inappropriate, harmful, or unsafe content generated by LLMs. It may involve human review or automatic moderation tools.

142
Q

Agentic workflow

A

A workflow that involves one or more AI agents carrying out tasks or decision-making processes. This may include coordination between multiple agents, or chaining multiple model actions for complex tasks (e.g., retrieve → analyze → generate → act).

143
Q

Prompt based learning

A

A method of training or using AI models where behavior is guided solely by the prompt, rather than altering the model’s internal parameters. Useful for adapting pre-trained models to new tasks without re-training.

144
Q

In context learning

A

The ability of LLMs to learn patterns and make predictions based solely on examples given in the prompt, without updating model weights. For instance, providing examples of Q&A in the prompt allows the model to generalize to a new question.

145
Q

Interactive prompting

A

The use of back-and-forth conversation (multiple turns) with an LLM to iteratively refine or clarify outputs. It’s common in chatbot interfaces or collaborative tasks.

146
Q

Shot prompting (zero, one shot few shot, n shot prompting)

A

A method that describes how many examples are included in the prompt to guide model behavior.

147
Q

Chain of thought prompting

A

A method where the prompt guides the model to generate intermediate reasoning steps, not just the final answer. Helps improve performance on complex reasoning tasks.

148
Q

Chain of reasoning prompt

A

Similar to chain-of-thought, but often used more formally to guide logical progression through substeps. Especially useful for solving multi-part or technical problems.

149
Q

Role prompting

A

Setting a context or role for the AI, like “You are a helpful tutor” or “Act like a data scientist.” This helps steer tone, perspective, and behavior in multi-turn interactions.

150
Q

Stochastic

A

Refers to randomness in a process. LLMs are stochastic models, meaning the same prompt may return different outputs due to probabilistic sampling (especially at higher temperature settings).

151
Q

Stochastic parrot

A

A term coined by researchers to critique LLMs as merely repeating patterns found in training data without true understanding—parroting information in a statistically likely way.

152
Q

Deterministic

A

Opposite of stochastic: a process that always gives the same output for the same input. In AI, setting temperature to 0 usually makes models behave deterministically.

153
Q

Emergent behavior

A

Unexpected capabilities that appear when LLMs reach a certain scale—e.g., performing math or translation even though they weren’t explicitly trained for it. These behaviors are not directly programmed but “emerge” from the model’s structure and training data.

154
Q

Architecture of an LLM

A

Refers to the underlying design and structure of a large language model. Most LLMs use a transformer architecture, with layers of attention mechanisms that process input sequences in parallel and understand context better than earlier models.

155
Q

Model fine tuning

A

The process of continuing to train a pre-trained AI model on a new, more specific dataset. This customizes the model for a particular use case (e.g., legal or medical language).

156
Q

Context window

A

The maximum number of tokens (words or symbols) a model can process at once. LLMs have limits (e.g., 4k, 16k, or 128k tokens), and exceeding the context window means earlier content may be truncated or lost.

157
Q

Stateless

A

Describes a system that doesn’t retain memory of previous interactions unless that context is explicitly provided again. Many LLMs operate statelessly unless memory is implemented at the application level.

158
Q

Reasoning models

A

AI systems (or LLMs configured in special ways) that are optimized to solve logical, mathematical, or multi-step problems by breaking down and reasoning through intermediate steps.

159
Q

Structured Query Language (SQL)

A

A standard language for accessing and manipulating relational databases. AI systems might generate or interpret SQL queries in data-centric applications.

160
Q

Python

A

A popular programming language widely used in AI and data science

161
Q

Multimodal models

A

AI models that can process and integrate multiple types of input data (modalities) such as text, images, audio, or video.

162
Q

AI winter

A

A historical period when interest and funding in AI drastically declined due to unmet expectations or lack of progress

163
Q

Fully connected ANN

A

A neural network where every neuron in one layer is connected to every neuron in the next layer. Common in early neural networks and some dense layers in deep learning models. Less efficient for image data due to high parameter counts.

164
Q

Loss function

A

A mathematical formula used during training to measure the difference between the AI model’s prediction and the actual (true) value. The model aims to minimize this loss to improve performance.

165
Q

Backpropogation

A

A key algorithm for training neural networks. It computes gradients of the loss function with respect to each weight in the network, allowing the model to adjust weights in the right direction using an optimizer like gradient descent.

166
Q

Edge detection (in AI vision models)

A

The process of identifying points in an image where brightness changes sharply—edges represent boundaries of objects. Edge detection is often an early layer function in vision models, enabling higher-level features like shape recognition.

167
Q

Convolution (processing of ai data)

A

A mathematical operation that slides a filter (kernel) over input data (e.g., an image) to extract localized features like edges or textures. It’s the core building block of Convolutional Neural Networks (CNNs).

168
Q

Kernel (in a convolutional neural network)

A

A small matrix (filter) used in the convolution operation to extract specific features from input data, such as edges, curves, or textures. Different kernels specialize in detecting different features.

169
Q

Feature (in a CNN)

A

A characteristic or pattern (e.g., edges, shapes, textures) detected by filters (kernels) at various layers. As you move deeper in the network, the features become more abstract (e.g., eyes, faces, objects).

170
Q

Feature detection

A

The process of identifying meaningful patterns in the input data. In CNNs, early layers detect low-level features (edges, gradients), while deeper layers detect complex structures (faces, animals, objects).

171
Q

Convolutional Neural Network

A

A deep learning architecture specialized for image and spatial data. It uses convolutional layers, pooling layers, and fully connected layers to automatically learn features and perform tasks like image classification, object detection, and segmentation.

172
Q

Image downsampling

A

The process of reducing an image’s resolution or size by removing pixels. This reduces computational load and highlights general patterns while removing fine details. Often used in CNNs via pooling layers.

173
Q

Image classification

A

A task in which an AI model assigns a label or category to an image (e.g., “cat”, “car”, “stop sign”). CNNs are commonly used for this task, trained on labeled datasets.

174
Q

CLIP (Contrastive Language-Image Pretraining)

A

An AI model developed by OpenAI that learns to match images with text descriptions. It’s trained to associate a caption with the correct image and distinguish it from mismatched ones, enabling tasks like zero-shot image classification or search.

175
Q

Generative adversarial network (GAN)

A

An architecture with two networks: a generator that creates synthetic data (e.g., images), and a discriminator that tries to detect whether data is real or fake. They compete in a “game” until the generator produces highly realistic data. GANs are behind many deep fakes and AI art.

176
Q

Diffusion models

A

A class of generative models that learn to generate data by reversing a process that gradually adds noise to input (diffusion). They start with noise and iteratively denoise to create realistic outputs—used in tools like DALL·E 2 and MidJourney.

177
Q

Image denoising

A

A process where AI removes visual noise from an image, often used in pre-processing or as a task for diffusion models. It enhances clarity and quality of visuals while preserving important features.

178
Q

Neural style transfer (NST) models

A

Models that recombine the style of one image (e.g., brush strokes of Van Gogh) with the content of another (e.g., a selfie). These models use CNNs to separate and recombine style and content layers of images.

179
Q

Text to speech (TTS)

A

Technology that converts written text into spoken audio using neural networks. Modern TTS systems, often powered by deep learning, can produce human-like voices and emotions.

180
Q

Deep fake

A

AI-generated or modified synthetic media—especially videos—where someone’s appearance or voice is altered to resemble someone else. Based on GANs and other generative models, deep fakes can be used for both creative and malicious purposes.

181
Q

NSFW (not safe for work)

A

A label for content that is inappropriate in professional or public settings, typically including violence, nudity, or explicit language. AI models must filter or block NSFW content when generating text or images to ensure safety and compliance.

182
Q

Hallucinations (in ai)

A

When an AI model generates factually incorrect, fabricated, or misleading information while sounding confident. This is a known issue in LLMs that stems from how they predict the next word statistically, not logically.

183
Q

Software access management

A

Processes and tools used to control who can access software systems, features, or data. Includes authentication (who you are) and authorization (what you can do), and is essential for securing AI tools and platforms.

184
Q

Software firewall

A

A security system that monitors and controls incoming and outgoing network traffic based on predetermined rules. It protects AI systems and servers from unauthorized access or malware.

185
Q

Data encryption

A

The process of encoding data so that only authorized parties can access it. AI applications use encryption to protect sensitive inputs/outputs and stored data from being intercepted or stolen.

186
Q

Secure socket layer (SSL)

A

An older protocol (now succeeded by TLS) used to encrypt communication between a user’s browser and a web server. It ensures that data sent to/from AI platforms is secure and private.

187
Q

Public key encryption

A

A type of encryption using a pair of keys: a public key to encrypt data and a private key to decrypt it. Used in secure communications for AI APIs, login credentials, and sensitive data transfers.

188
Q

Packet sniffing

A

A technique where network traffic is monitored to capture data packets. Malicious actors may use sniffing to steal unencrypted data; AI systems need protection via encryption and secure protocols.

189
Q

Software logging

A

The process of recording events, errors, and user actions in a system. In AI, logs are crucial for debugging, monitoring usage, detecting attacks, and ensuring accountability.

190
Q

Intellectual Property (IP)

A

Creations of the mind—such as inventions, designs, symbols, or software—that are legally protected. In AI, IP can include code, trained models, datasets, and even prompts or generated content.

191
Q

IP Leakage

A

When proprietary or confidential information is unintentionally exposed or inferred from AI model outputs. Can happen through LLMs trained on sensitive data or careless system configurations.

192
Q

Patents

A

Legal protections granted to inventors, allowing exclusive rights to an invention for a limited time. Patents can cover AI algorithms, systems, or specific technical solutions.

193
Q

Patent novelty

A

A requirement that the invention must be new, not publicly disclosed or used before the filing date. If a method has already been shared (even on GitHub), it may not be patentable.

194
Q

Patent non obviousness

A

The invention must not be an obvious solution to someone with standard skills in the field. In AI, combining existing models in a slightly different way might not qualify unless it solves a problem in a novel way.

195
Q

Patent utility

A

The invention must be useful and have a practical application. This is usually easy to demonstrate for software and AI inventions that perform real tasks or improve efficiency.

196
Q

Trade secrets

A

Proprietary processes, data, or methods kept confidential to maintain a competitive edge. Unlike patents, trade secrets aren’t registered but must be protected from disclosure or theft (e.g., Google’s search algorithm).

197
Q

Public domain

A

Refers to content or inventions that are not protected by IP laws and can be freely used by anyone. AI-generated work may or may not qualify for IP protection, depending on the jurisdiction.

198
Q

AI Model transparency

A

The degree to which the workings of an AI model can be explained, understood, or inspected. Transparency is key to ethical AI, especially in high-risk domains like healthcare or finance.

199
Q

LLM Filters

A

Algorithms or rule-based systems that restrict or modify the inputs/outputs of large language models to prevent harmful or unwanted content (e.g., profanity, misinformation, NSFW material).

200
Q

LLM input/output moderation

A

Mechanisms for reviewing and managing the prompts users send to a model and the responses it returns. This may include auto-moderation tools or human oversight.

201
Q

Data poisoning

A

A type of attack where malicious data is inserted into the model’s training set to corrupt or bias its behavior. For example, inserting harmful associations into a chatbot’s training data.

202
Q

Adversarial perturbations

A

Tiny, carefully designed changes to input data (like an image) that cause a model to misclassify or behave incorrectly—even if humans don’t notice the difference. A key threat in AI vision systems.

203
Q

Adversarial training

A

A defense technique where models are trained on adversarial examples to become more robust against these attacks. Often used in computer vision and cybersecurity-sensitive applications.

204
Q

Black hat hacker

A

A malicious actor who exploits vulnerabilities in software or AI systems for unethical or illegal purposes, such as stealing data, manipulating outcomes, or spreading misinformation.

205
Q

Data privacy

A

The right and practice of ensuring that individuals’ personal data is collected, processed, and shared in a secure and lawful manner. In AI, data privacy is vital when using user-generated content, medical records, or personal profiles for training or inference.

206
Q

PII (Personally identifiable information) is

A

Any information that can identify an individual, such as names, addresses, email addresses, social security numbers, and biometric data. AI systems that process PII must comply with regulations like GDPR or CCPA to prevent misuse or exposure.

207
Q

Copyright law

A

Legal protection granted to creators of original works, such as text, images, code, music, and more. Copyright gives authors exclusive rights to reproduce, distribute, and adapt their work, and is highly relevant in AI-generated content debates.

208
Q

Fixation (in copyright law)

A

For a work to be protected by copyright, it must be “fixed in a tangible medium” (e.g., written, recorded, saved digitally). Transient ideas or spoken words not recorded typically don’t qualify.

209
Q

Originality (in copyright law)

A

A copyrightable work must show a minimal degree of creativity and be independently created. AI-generated content that mimics existing works too closely may fail this standard.

210
Q

Idea expression dichotomy

A

A legal principle stating that ideas are not protected by copyright—only the expression of those ideas is. This means anyone can use the idea behind a copyrighted work but not copy how it’s expressed.

211
Q

Derivative work

A

A new creation that is based on or adapted from an existing copyrighted work (e.g., a remix, translation, or fan fiction). These works may still be subject to the original copyright holder’s permission.

212
Q

Common Crawl

A

A non-profit organization that regularly crawls and stores petabytes of public web data. It’s used by many AI companies (e.g., OpenAI, Meta) to train language models, though its inclusion of copyrighted content has raised legal and ethical concerns.

213
Q

Fair use (in copyright law)

A

A doctrine allowing limited use of copyrighted works without permission for purposes like criticism, commentary, teaching, and research. In AI, fair use is a legal gray area, especially when models are trained on copyrighted data.

214
Q

Open source licensing

A

Legal frameworks that allow software to be used, modified, and distributed freely under specific conditions. Examples include MIT, Apache 2.0, and GNU GPL. Open source fosters collaboration, but AI developers must understand the terms.

215
Q

Open Responsible AI License (OpenRAIL)

A

A newer license type developed to promote ethical use of open-source AI models. It allows sharing and usage while placing restrictions on harmful use cases (e.g., surveillance, weaponization).

216
Q

Creative commons (CC)

A

A set of licenses that enable creators to specify how others can use their work (e.g., share, remix, or use commercially). Ranges from CC-BY (attribution only) to CC0 (public domain dedication).

217
Q

Copyleft

A

A licensing approach that allows derivative works but requires them to remain under the same license. Ensures that modified versions of software or models remain open and free. Example: GNU General Public License (GPL).

218
Q

Indemnification

A

A legal agreement where one party agrees to compensate another for harm or liability. In AI contracts, software providers may offer indemnification if their tool causes legal trouble (e.g., IP infringement or data breach).

219
Q

Liability

A

Legal responsibility for harm or damages caused by an AI system or its outputs. Companies deploying AI must assess whether liability falls on the developers, users, or integrators—especially when AI makes autonomous decisions.

220
Q

Governance policies

A

Internal or organizational rules that ensure the ethical, legal, and responsible use of AI systems. These cover data use, bias mitigation, transparency, auditing, access control, and adherence to AI principles.

221
Q

AI Fairness

A

Ensuring AI systems do not produce biased or discriminatory outcomes against any group based on race, gender, age, or other attributes. Fairness is a key consideration in training data, model evaluation, and deployment.

222
Q

AI Model Alignment

A

Ensuring that an AI model’s behavior aligns with human values, goals, and expectations. Misalignment occurs when models act in ways that are harmful, manipulative, or unexpected—even if technically correct.

223
Q

False positive

A

Occurs when an AI system incorrectly flags something as inappropriate, dangerous, or incorrect (e.g., marking harmless content as toxic). Important to minimize in moderation systems.

224
Q

Prompt hacking

A

Tricking an AI model into ignoring safety constraints or providing unintended responses by crafting misleading or manipulative prompts. Related to jailbreak attacks.

225
Q

AI Model jailbreaking

A

A more advanced version of prompt hacking where users bypass built-in restrictions of a model to make it behave in unsafe, unethical, or unauthorized ways.

226
Q

Constitutional AI

A

An approach where models are guided by a “constitution”—a set of ethical rules or principles that shape how they behave. Introduced by Anthropic, it’s an attempt to enforce value alignment through self-feedback rather than human oversight alone.

227
Q

Explainable AI

A

A set of techniques and tools that help humans understand how and why an AI system made a specific decision or prediction. Critical for trust, adoption, and compliance in regulated industries.

228
Q

Counterfactual explanations

A

A form of XAI where the system explains how an output would have changed if the input had been slightly different. Example: “If your income had been $5,000 higher, your loan would have been approved.”

229
Q

Local interpretable model agnostic explanations (LIME)

A

An XAI technique that explains a model’s prediction by approximating it with a simpler, interpretable model near the prediction. It works with any black-box model and is used for debugging and compliance.

230
Q

Attention mechanisms

A

A key innovation in transformer-based models like GPT. It allows the model to “attend” to relevant parts of the input when making predictions. Attention helps models handle long-range dependencies and understand context.

231
Q

LLM Ensemble approach

A

Combining multiple LLMs or variations of a model to improve reliability, reduce hallucinations, or increase robustness. The ensemble may vote, average responses, or select the best output from multiple generations.

232
Q

Model chaining

A

Connecting multiple AI models in sequence, where the output of one model becomes the input to another. Useful for multi-step reasoning, multi-modal generation, or agentic workflows.

233
Q

Long tail distribution

A

A statistical pattern where a small number of items (the “head”) are very popular, while a vast number of items (the “tail”) each have low frequency or demand. A few jobs make the majority of the money, while many jobs make little to no money

234
Q

Marginal cost

A

The cost of producing one additional unit of a product or service. In AI, the marginal cost of using a model (once trained) is often near zero—running an extra query costs very little, especially with scalable cloud infrastructure.

235
Q

Corpus

A

A large and structured collection of texts or data used to train language models. Common corpora include books, websites (like Common Crawl), news articles, and research papers. The quality, diversity, and bias of the corpus greatly affect model behavior.

236
Q

Internet of Things (IoT)

A

A network of physical devices (e.g., thermostats, fridges, vehicles, sensors) connected to the internet, able to collect and share data. When integrated with AI, IoT enables smart automation, predictive maintenance, and personalized services.

237
Q

Universal basic income

A

An economic proposal where all citizens receive a regular, unconditional payment from the government. UBI is often discussed in the context of AI-driven job automation and economic disruption as a potential solution for income stability.

238
Q

AI Echo Chamber

A

Occurs when AI systems, especially recommendation engines or chatbots, reinforce users’ existing beliefs or biases by repeatedly presenting similar views or responses. This can lead to information bubbles and polarization.

239
Q

Superintelligent AI

A

A hypothetical AI that surpasses human intelligence across all domains—logic, creativity, emotional intelligence, and more. It’s the subject of much debate in AI safety and philosophy due to its potential to reshape civilization or pose existential risks.

240
Q

AI singularity

A

A theorized future point when AI systems achieve recursive self-improvement, rapidly evolving beyond human understanding or control. It marks the potential moment of irreversible change in human society, sometimes seen as either utopian or catastrophic.

241
Q

Utopian view of ai

A

A positive outlook where AI improves lives, eliminates tedious labor, cures diseases, helps solve climate change, and creates a fairer, more abundant society. Often tied to visions of harmonious human-AI collaboration.

242
Q

Dystopian view of ai

A

A negative projection where AI leads to mass surveillance, authoritarian control, mass unemployment, misinformation, or loss of human autonomy. These fears are fueled by unchecked deployment, misalignment, or misuse of powerful models.