Domain 2 Flashcards
What is generative AI?
Generative AI is a subset of deep learning. Like deep learning, generative AI is a multipurpose technology that helps to generate new original content rather than finding or classifying existing content. Generative AI focuses on creating new content, such as text, images, audio, video, and even code
Gen AI foundational models have small numbers of parameters
False. very large and complex neural network models with billions of parameters that are learned during the training phase or pre-training
What is the current core element of Gen AI?
Transformer network. Transformers were introduced in a 2017 paper called “Attention Is All You Need.” Some LLMs, such as ChatGPT, are built on the transformer architecture. These LLMs are pre-trained on massive amounts of the text data from the internet. They can use this pre-training process to build up a broad knowledge base. And they can be fine-tuned for specific tasks with relatively little additional data.
Prompt
The input that you sent into your generative model is called the prompt.
inference
s the process that a trained machine learning model* uses to draw conclusions from brand-new data. An AI model capable of making inferences can do so without examples of the desired result.
Completion
Context window
Tokens
a token is a fundamental unit of data that is processed by algorithms, especially in natural language processing (NLP) and machine learning services. A token is essentially a component of a larger data set, which may represent words, characters, or phrases.
LLMs’ vocabulary
Tokenizer
Before your text is sent to the AI, it gets turned into numbers in a process called tokenization. These tokens are how the AI reads and interprets text.
Prompt engineering
is the process of creating and refining inputs, or prompts, for AI models to produce the desired outputs. It’s a complex process that involves providing context, instructions, and examples to guide the AI model to understand the user’s intent and respond in a meaningful way.
What is in-context learning?
to get the model to produce better completions is to include examples of the task that you want the model to carry out. These examples can be incorporated inside the prompt. You can use few-shot, zero-shot, and one-shot inference.
These influence the model’s completion to the prompt
Inference configuration parameters
every language-based generative AI model has this, which converts human text into a vector that contains token IDs or input IDs. Each input ID represents a token in the model’s vocabulary.
tokenizer
What is a vector?
A vector is an ordered list of numbers that represent features or attributes of some entity or concept. In the context of generative AI, vectors might represent words, phrases, sentences, or other units. The power of vector representations is in the ability to encode related relationships between the items and capture meaningful associations, analogies, and hierarchies.
Embedding vectors are also called embeddings.
Embeddings are a numerical vectorized representation of any entity. Embeddings capture the semantic meaning of tokens such as text, image, video, or audio. For example, the vectors encode the meaning and context of tokens within a large body of text.
Name an innovation of transformers
An innovation of transformers is this self-attention mechanism. This mechanism helps the model to weigh the importance of different parts of the input when generating each output token. As a result, the model can capture long-range dependencies and contextual relationships that were difficult to learn with previous architectures
What are position embeddings?
Transformers also introduce the concept of position embeddings, which encode the relative position of each token in the sequence. They help the model to distinguish between identical tokens that appear in different positions, which is important for understanding sentence structure and word order.
encoder
generates an embedding or vector representation for each token
Self-attestation
Self-attestation is a process where an organization or individual declares that they comply with a specific set of rules or standards without the need for third-party verification. Self-attestation is often used in the context of cybersecurity and compliance, but it can also be used in other contexts.
decoder
also known as generators, are responsible for translating these latent vectors back into meaningful output data. They reconstruct the data based on the learned patterns and relationships from the encoded space, resulting in outputs that often exhibi
Soft max output
the softmax output is a vector of probabilities that represent the likelihood of each class label in a multi-class classification problem. The softmax function is an activation function that is often used in the final layer of a neural network model
What percentage of data is used for pre-training after the data quality curation step?
1% to 3%
Researchers have found that the larger a model is…
the more likely it is to work without additional in-context learning or further training. Because the model’s capability increases with size, it has supported the development of larger and larger models.
Name the two kinds of Generative AI
Unimodal and multimodal
What is unimodal Gen AI
Unimodal models work with one data modality. LLMs are an example of unimodal generative AI because the input and the output, or completion, are text.
What is multimodal Gen AI
Multimodal is adding another modality such as image, video, or audio. Multimodal models can understand diverse data sources and can provide more robust forecasts. Multimodal generative AI use cases are marketing, image captioning, product design, customer service, chatbots, and avatars
What are two important classes that go beyond text only applications?
Multimodal and diffusion models
What are some examples of multimodal tasks?
These are image captioning, where the model is generating text descriptions of images, visual question answering, where the model answers questions about image content. Another example is text to image synthesis, which is generating images from textual descriptions
What are some examples of models that produces realistic and diverse images?
DALL-E, Stable Diffusion, and Midjourney
These models are a class of generative models that learn to reverse a gradual noising process.
Diffusion models. Diffusion-based architectures offer a higher degree of control in quality and diversity of images generated
forward diffusion
The forward diffusion process begins by sampling from a basic, usually Gaussian, distribution. This initial simple sample undergoes a series of reversible, incremental modifications, where each step introduces a controlled amount of complexity through a Markov chain.
reverse diffusion
is the opposite of the forward diffusion process, where a model learns to recover original data from noisy data. The process involves training a model to find reverse Markov transitions that maximize the likelihood of the training data
Stable diffusion
a generative artificial intelligence (generative AI) model that produces unique photorealistic images from text and image prompts. It
T or F: Diffusion models tend to produce lower quality outputs with less diversity and consistency, and they’re less stable and harder to train.
Diffusion models tend to produce higher quality outputs with more diversity and consistency, and they’re more stable and easier to train.
What are some use cases for Generative AI?
writing or rewriting pieces of text to adapt to different audiences
text summarization
code generation and completion
information extraction, question answering, classification, identifying harmful content, translation, recommendation engines, personalized marketing and ads, chatbots, customer service agents, and search
What is the rebranded name of Amazon CodeWhisper?
Amazon Q Developer
What Gen AI services help with virtual production and 3D content creation?
Amazon Nimble Studio and Amazon Sumerian
generative adversarial networks
Given a training set, this technique learns to generate new data with the same statistics as the training set. For example, a GAN trained on photographs can generate new photographs that look at least superficially authentic to human observers, having many realistic characteristics. Though originally proposed as a form of generative model for unsupervised learning, GANs have also proved useful for semi-supervised learning,[4] fully supervised learning,[5] and reinforcement learning.[6]
variational autoencoders
artificial neural network architecture introduced by Diederik P. Kingma and Max Welling.[1] It is part of the families of probabilistic graphical models and variational Bayesian methods.[2]
In addition to being seen as an autoencoder neural network architecture, variational autoencoders can also be studied within the mathematical formulation of variational Bayesian methods, connecting a neural encoder network to its decoder through a probabilistic latent space (for example, as a multivariate Gaussian distribution) that corresponds to the parameters of a variational distribution.
What are the stages in the Gen AI project life cycle?
identify use case; experiment and select;
adapt, align, and augment;
evaluate;
deploy and iterate;
monitor
The most important step in any project is to define the scope as accurately and narrowly as you can.
You should think about what function the LLM will have in your specific application. Do you need the model to be able to carry out many different tasks, including long-form text generation, or is the task much more specific, like named entity recognition, so that your model only needs to be good at one thing. Getting specific about what you need your model to do, You can save time and perhaps more importantly, compute costs.
What are the steps in the foundation model lifecycle?
data selection,
model selection (foundational or build your own)
pre-training,
fine-tuning, evaluation,
guardrails
deployment
monitoring
feedback
What is an additional fine-tuning technique?
Reinforcement learning (from human feedback)
Re-enforcement learning
a machine learning technique that teaches software how to make decisions to achieve the best outcomes. It’s based on the idea that the best way to learn is through trial and error, and it’s often used in robotics and gaming. RL algorithms learn by interacting with an environment and observing how it responds. They use a reward-and-punishment system to reinforce actions that help them achieve their goals, and ignore actions that don’t.
What are three advantages to generative AI?
adaptability
responsiveness
simplicity
Do LLM’s learn from interacting with you?
Every time you prompt your LLM, the LLM does not actually remember earlier conversations. It is similar to asking a different child for every single task. Therefore, you don’t get to train them over time on specifics of your business or the style you want them to write, but you could with fine-tuning
What is Fine Tuning?
Fine tuning is an alternate approach to GenAI development that involves training an LLM on a smaller, specialized, labeled dataset and adjusting the model’s parameters and embeddings based on new data. This is different from RAG.
What are two methods for model interpretability?
Intrinsic Analysis
Post Hoc Analysis
ROUGE, or Recall-Oriented Understudy for Gisting Evaluation,
is primarily employed to assess the quality of automatically-generated summaries by comparing them to human-generated reference summaries
BLEU, or Bilingual Evaluation Understudy,
is an algorithm designed to evaluate the quality of machine-translated texts by comparing it to human-generated translations.
autoregressive models.
a class of machine learning (ML) models that automatically predict the next component in a sequence by taking measurements from previous inputs in the sequence.
By analyzing large amounts of business data to forecast their future values or to detect outliers and understand the root cause is complex, time consuming, and not always accurate. AWS provides Amazon’s business metric analysis ML solution which uses Amazon Lookout for Metrics and Amazon Forecast to solve these problems. It uses machine learning to analyze large volumes of data while dynamically adapting to changing business requirements.
Organizations need to evaluate potential return on investment and weighing the cost and benefits of FMs considering their application. Additionally, it’s important to understand the metrics for comparing operational costs
What are the advantages of using OOTB services from AWS for Gen AI?
Accessibility
Lower barrier to entry
Efficiency
Cost-effectiveness
Speed to market
Ability to meet business objectives
AWS Nitro
Security layer
Three critical components of AI systems that need to be secured
Input
model
output
Give some examples of AI system vulnerabilites
prompt injection, data poisoning, and model inversion vulnerabilities
Two pricing models for LLMS
Host your own infra
token-based pricing
AWS ML Stack
AWS AI Services (APIs, SDKs)
AWS ML Services
Amazon SageMaker
Amazon Infrastructure, EC2, GPUs, and more
SageMaker Jump Start
Pretrained models
Find tune your data
Deploy using SageMaker
Amazon Bedrock
Gives you the ability to interact with different best in class models, including its own models (Titan). mazon Bedrock adds the capability to import custom weights for supportive model architectures, and serve the custom model by using on-demand mode
What is Amazon Bedrock’s Playground?
Playgrounds in Amazon Bedrock let you experiment by running model inference against different base foundation models that are supported within the service to help you align your use cases with the highest accuracy. Depending on the model selected for your playground, it will determine the right types of inference parameters that you can adjust. And remember, you can vary the number of inference parameters to determine different completion results.
Remember that with generative AI, you can use vector databases, and that data is stored as embeddings. These embeddings are vectors that can be compressed, stored and indexed for advanced searches