AWS AI Practitioner Flashcards
Portion of training training data is labeled and feedback is provided in the form of rewards or penalties. What type of learning
Reinforcement learning
What are the two types of inferencing?
Batch and Real time
Deep learning is used in which use 2 cases?
Computer vision and NLP
What are FMs in generative AI?
Pretrained models
What are Transformer models?
Builds encoder decoder concept in genAI. They use self-attention to process input data. Self-attention allows the model to weigh the importance of different words in a sentence when encoding a particular word
Are FMs pre trained using reinforced learning? True or False?
False. FMs are typically pre-trained through self-supervised learning
Where is pre-text tasks used for?
In self-supervised learning
Self-supervised learning makes use of the structure within the data to autogenerate labels. True or False?
True
Optimization of pre trained FMs are done using what?
Prompt engineering,
Retrieval-augmented generation (RAG),
Fine-tuning on task-specific data
LLMs, Diffusion and Multiodel models are what?
These are FM models
These are numerical representations of tokens, where each token is assigned a vector (a list of numbers) that captures its meaning and relationships with other tokens?
Embeddings
What is a context window?
The maximum number of tokens a LLM model can take when generating text
What is a vector?
It is an array of numercial values
What is the process of vectorization?
Text -> [Tokenization]->Tokens -> [Embeddings Model] -> Vectors
What is the process of vectorization in Bedrock KBs using RAG?
Customer KB->[Upload in Amazon S3]->[Select a vector DB]->[Select a Model]->[Sync with customer KB]->Vectorization of Customer KB text
What is Watermark detection for Amazon Bedrock?
Identify images generated by Amazon Titan Image Generator, a foundation model that allows users to create realistic, studio-quality images in large volumes and at low cost, using natural language prompts
What is continued pretraining in Amazon Bedrock?
You provide unlabeled data to pre-train a model by familiarizing it with certain types of inputs
Which is the models which gradually add more and more meaningful information to this noise until they end up with a clear and coherent output, like an image or a piece of text?
Diffusion model
Which model has generator and discriminator?
Generative adversarial networks
Which model has encoders and decoders?
Varional autoencoders
What are the components of prompt engineering
Instructions, Context, Input data and Output indicator
What are non-determistic LLMs popularly called?
Generative Language Models
What is a supervised learning process that involves taking a pre-trained model and adding specific, smaller datasets?
Fine tuning
Two types of fine tuning
Instruction fine-tuning and Reinforcement learning from human feedback (RLHF)
Fine tuning does it add weight to the data?
Yes
What is Retrieval-augmented generation (RAG)?
Supplies domain-relevant data as context to produce responses based on that data.
To create fine tuned models in Bedrock what is the pricing option?
Provisioned Throughput only which billed by the hour
How is RAG different from fine tuning
Rather than having to fine-tune an FM with a small set of labeled examples, RAG retrieves a small set of relevant documents and uses that to provide context to answer the user prompt
What are two types of supervised learning?
Classification and Regression
Predicting continuous or numerical values based on one or more input variable?
Regression
Forcasting uses which supervised learning technique?
Regression
Diagnostic uses which supervised learning technique?
Classfication
What are two types of unsupervised learning?
Clustering and Dimensionality reduction
Examples of RAG vector databases?
Amazon OpenSearch Service(KNN capability, vector embeddings),
DynamoDB(high performance,vector embeddings),
Aurora(RDS),
RDS for PostgreqSQL(RDS and open source),
Neptune(GraphQL)
Grouping of unstructured data is done in which type of unsupervised learning?
Clustering
Reducing the number of features or dimensions in a dataset in which type of unsupervised learning?
Dimensionality reduction
Which learning type continuously improves its model by mining feedback from previous iterations?
Reinforcement learning
In which learning the reward of a desired outcome is known, but the path to achieving it isn’t?
Reinforcement learning
How to reduce toxity risk in generative AI?
Use guardrail models
What does guradrail models do?
These models will detect and filter out unwanted content
What is the risk term for when model generates inaccurate responses that are not consistent with the training data?
Hellucinations
What is the risk term when model might generate different outputs for the same input?
Nondeterminism
What is the risk term when the information shared with your model can include personal information and can potentially violate privacy laws?
Data security and privacy concerns
What is the risk term when output generated by model has PII?
Regulatory violations
Which generative AI model used for chatbots?
Llama
Which generative AI model used for code generation?
Claude
Which generative AI model used for code gaming?
Stable Diffusion
Which generative AI has embeddings?
Amazon Titan
Which generative AI has a use case of Healthcare – summarize key ideas from long text?
Command
What are the capabilities of generative AI?
SPARCD
Adaptability
Responsiveness
Simplicity
Creativity and exploration
Data efficiency
Personalization
Recommendation engines, gaming, and voice assistance are examples of which type of AI system?
Traditional AI
Chatbots, code generation, and text and image generation are examples of which type of AI system?
Generative AI
When is a model is underfitted?
When a model has a high bias
Overfitting happens when?
When model performs well on the training data but does not perform well on the evaluation data
SageMaker Clarify
You can automatically evaluate FMs for your generative AI use case with metrics such as accuracy, robustness, and toxicity to support your responsible AI initiative
SageMaker Clarify is used for text based models only. True or False?
True
Model evaluation on Amazon Bedrock
Evaluate, compare, and select the best foundation model for your use case in just a few clicks
Can in human evaluation, we can automate it?
Yes, using built in task types
Amazon Bedrock Guardrails
Guardrails helps control the interaction between users and FMs by filtering undesirable and harmful content, redacting personally identifiable information (PII), and enhancing content safety and privacy in generative AI applications
Amazon SageMaker Data Wrangler
Offers three balancing operators: random undersampling, random oversampling, and Synthetic Minority Oversampling Technique (SMOTE) to rebalance data in your unbalanced datasets
Amazon SageMaker Experiments
Provide scores detailing which features contributed the most to your model prediction on a particular input for tabular, natural language processing (NLP), and computer vision models
You can use to create, manage, analyze, and compare your machine learning experiments.
Amazon A2I
Human review of ML predictions
SageMake governance tools
Amazon SageMaker Role Manager - define minimum permissions in minutes
Amazon SageMaker Model Cards - capture, retrieve, and share essential model information, such as intended uses, risk ratings, and training details, from conception to deployment
Amazon SageMaker Model Dashboard - You can keep your team informed on model behavior in production, all in one place
AWS AI Service Cards
Responsible AI documentation
Basic concepts to help customers better understand the service or service features
Intended use cases and limitations
Responsible AI design considerations
Guidance on deployment and performance optimization
What are the four elements of responsible agency in responsible AI?
Value alignment
Responsible reasoning skills
Appropriate level of autonomy
Transparency and accountability
What is Data curation?
Curating datasets is the process of labeling, organizing, and preprocessing the data
What are the three steps of data curation?
Data preprocessing, augmentation and audit
Whar are the AWS tools for transparency?
AWS AI Service Cards and Amazon SageMaker Model Cards
Whar are the AWS tools for explainability?
SageMaker Clarify and SageMaker Autopilot
What to use to build your custom ML model in SageMaker with less code?
Sagemaker Canvas using AutoML powered by Sagemaker Autopilot
Which is more detailed interpretability or explainability ?
Interpretability
PDP graphs are used for what?
They tell how single feature influence the predicted outcome. Used for interpretibiity and explainability
What are human centered design concepts?
Design for amplified decision making
Design for unbiased decision making
Design for human and AI learning
This principle seeks to maximize the benefits of using technology while minimizing potential risks and errors, especially risks and errors that can occur when humans make decisions under stress or in high-pressure environments
Design for amplified decision-making
The design of decision-making processes, systems, and tools is free from biases that can influence the outcomes
Design for unbiased decision-making
This priciple aims to create learning environments and tools that are beneficial and effective for both humans and AI
Design for human and AI learning
Reinforcement learning from human feedback
(RLHF) is an ML technique that uses human feedback to optimize ML models to self-learn more efficiently
Which AWS tool provide RLHF?
Amazon SageMaker Ground Truth
What is Feature engineering?
It is the process of creating, transforming, extracting, and selecting variables from data. Convert raw data into meaningful data
Common ratio for data training, validation and testing?
80,10,10 or 70,15,15
Amazon Sagemake which part does “Collecting, analyzing, and preparing your data”?
Amazon SageMaker Data Wrangler
Amazon Sagemake which part does “Managing Features”?
Amazon SageMaker Feature Store
Amazon Sagemake which part does “Model training and evaluation”
Amazon SageMaker Canvas
Amazon SageMaker Canvas
Access to ready models from Bedrock and Jumpstart and No coding is required
It is integrated with Comprehend, Rekognition and Textextract
Amazon SageMaker JumpStart?
Provides pretrained, open source models that customers can use for a wide range of problem types
Amazon Sagemaker which part does “Model evaluation”
Amazon SageMaker Experiments
Amazon Sagemaker which part does “Hyperparameter tuning”
Amazon SageMaker Automatic Model Tuning
What is learning rate hyperparameter?
So if you have a higher learning rate, that means that your model is going
to have a faster conversions, but there is a risk of you to overshoot the optimal solution because while you’re going too fast for learning.
And if you have a low learning rate, it may be more precise but slower convergence
Amazon Sakemaker which part does “Monitoring”
Amazon SageMaker Model Monitor
Supervised learning Sagemaker built in algorithms
(FxKL)
Linera learner
Factorization machines
XGBoost
K-Nearst Neighbours(KNN)
Unsupervised learning Sagemaker built in algorithms
Clustering - K-means, LDA
Topic Modeling - LDA
Embeddings - Object2Vec
Anomoly detection - Random cut forest, IP insights
Dimensionality reduction - Pricipal component analysis (PCA)
Image processing sagemake built in algorithms
Image classification - MXNet tensor flow
Object detection - MXNet tensor flow
Semantic segmentation - FCN,PSP,Deeplab V3
Time series - DeepAR
Text Analysis sagemake built in algorithms
Text classification -Blazing text
Word2Vec - Blazing text
Machine translation - Sequence to sequence
Topic modeling - LDA,NTM
Speech - Sequence to sequence
When there is greater gap between predicted and actual value what it means?
High bias
When the predicted values are too much dispersed what it means?
High variance
How to reduce high variance?
Feature selection for more important features and multiple sets of training and test sets of data
What matrix help classify why and how a model gets something wrong?
Confusion matrix
Confusion matrix?
It is used to evaluate the performance of the model that does classfication
Formula for accuracy
(TP+TN)/(TP+FP+TN+FN)
Formula for Precision
(TP)/(TP+FP)
When to use precision over accuracy?
When the cost of false positives are high in your particular business situation
Think about a classification model that identifies emails as spam or not. In this case, you do not want your model labeling a legitimate email as spam and preventing your users from seeing that email.
Formula for Recall
(TP)/(TP+FN)
When to use recall over precision?
If it is extremely important and vital to the success of the model that it not give false negative results
Think about a model that needs to predict whether a patient has a terminal illness or not
Formula for AUC-ROC
ROC is a probability curve, and AUC represents the degree or measure of separability.
In general, AUC-ROC can show what the curve for true positive compared to false positive looks like at various thresholds.
What is Mean squared error?
You take the difference between the prediction and actual value, square that difference, and then sum up all the squared differences for all the observations and divide by number of predictions
What is R squared?
R squared explains the fraction of variance accounted for by the model
What is the difference between Mean squared error and R squared?
MSE focuses the measure of model performance.
R squared provides a measure of the model’s goodness of fit to the data.
What is A/B testing or the canary deployments technique?
Developers can experiment with two or more variants of a model and help achieve the business goals.
For real time interactive workloads with low latency requirements, which deployment model?
Real-time
Requests with large payload sizes(upto 1GB), high processing time and near real time latency requirement, which deployment model?
Asynchronous
Large dateset and don’t need a persistent endpoint, which deployment model?
Batch transform
Workloads that have idle periods and can tolerate cold starts, which deployment model?
Serverless
Providing self-service enviornments and curated data sets, which benfit of MLOps?
Productivity
Incoporating CI/CD practices, which benfit of MLOps?
Reliability
MlFlow
Manage entire ML lifecycle in Sagemaker Studio
Automating all steps in ML development lifecycle, which benfit of MLOps?
Repeatibility
Versioning all inputs and outputs , which benfit of MLOps?
Auditibility
Policies to guard against model bias and track changes to data statistical properties, which benfit of MLOps?
Quality
Sagemaker pipeline
Prepare data:
Sagemake data wrangle
Sagemake processing job
Curate feature:
Sagemake feature store
Experiment tracking:
Sagemaker experiments
Train model:
Sagemaker training job
Evaluate model:
Sagemaker processing job
Register model:
Sagemaker model registry
Deploy model:
Deployments
Manage model:
Sagemaker model monitor
GenerativeAI pretrained models - J2
LLM - text generation, contextual question answering, summarization, and classification
GenerativeAI pretrained models - Amazon Titan
Embeddings, text generation, and image generation
GenerativeAI pretrained model - Claude
Art vision and text AI models
GenerativeAI pretrained model - Command XL
Text-based responses based on prompts
GenerativeAI pretrained models -Llama 3
LLM - generate coherent and contextually relevant text
GenerativeAI pretrained models -Mistral Large
large reasoning capabilities or are highly specialized, like synthetic text generation, code generation, RAG, or agents
GenerativeAI pretrained models -Stable Diffusion
Can generate images of from text input.
How to increase the performace of the model?
Using prompt engineering, RAG, fine-tuning, or automation agents
Prompt engineer works for which models?
LLMs
What is demonstrations, or task-specific instructions called in prompt engineering?
Augmentation
What is Iteratively refining and adjusting the prompts called?
Refining
What is combining multiple prompts or generation strategies?
Ensembling
What is prompt searching, prompt generation, or prompt retrieval from large prompt libraries?
Mining
Three use cases of RAG
- Customer support, virtual assistants
- Journalism and research
- Content marketing
What is Amazon Bedrock knowledge base?
Provide you the capability of amassing data sources into a repository of information
What model optmization technique used by Amazon Bedrock knowledge bases?
RAG without customization
Pricing of Amazon Bedrock
Pay as you go
Based on no of tokens in input and response for text based models
Pay as you go
Based on no of images in input and response for image based models
Batch
Multiple predictions at a time and sent as 1 file to S3
50% discount
Provisioned Throughput
Based on no of input and response tokens processed each minute for text based models and is called Provisioned Throughput
What is Fine-tuning?
Taking a pre-trained language model and further training it on a specific task or domain-specific dataset
What is another name of instruction based fine tuning?
Prompt Tuning
Instruction based find tuning features
Labeled examples
Prompt-response pairs
Continued Pre-training features
Particular field or area of knowledge
Unlabeled data
Domain specific training
What is another name of human feedback data used in fine tuning?
Reinforcement learning from human feedback (RLHF)
Which type of learning creates a reward model?
RLHF
Steps in creating a model from scratch?
- Selecting the appropriate neural network architecture, layers, and hyperparameters
- A large and diverse dataset must be curated, cleaned, and preprocessed
- Model is initialized with random weights and trained using various optimization algorithms
What are agents used during performance improvement of a model?
Carry out various multi-step tasks related to infrastructure provisioning, application deployment, and operational activities.
Task coordination
Reporting and logging
Scalability and concurrency
Integration and communication
Model improvement techniques and pricing impact
Cheapest to Costly
Prompt Engineering - No model training needed
RAG - Use external knowledge but no FM changes. Cost for using vector dbs
Instruction based fine tuning - FM is fine tuned with instructions and change the tone of the model.
Domain Adaption fine tuning - Domain specific model training
Does changing temperature, Top K, Top P affects model pricing?
No
Does temperature, Top K, Top P affects model latency?
No
Types of benchmark datasets for generativeAI evaluation?
GLUE - text classification, question answering, and natural language inference
SuperGLUE - compositional language understanding
SQuAD - question-answering capabilities
WMT - machine translation systems
What are three types of automated metrics?
Perplexity (a measure of how well the model predicts the next token)
BLEU score (for evaluating machine translation)
F1 score (for evaluating classification or entity recognition tasks)
Formula for F1?
(2xPrecisionxRecall) / (Precision+Recall)
What is the advantage of automated metrics?
Automated metrics can be useful for rapid iterations and fine-tuning during model development
What are the disadvatage of automated metrics?
They fail to capture the nuances and complexities of human language and might not align perfectly with human judgments
ROUGE automated metrics definition?
Automatic summarization and machine translation systems
The main idea behind ROUGE is to count the number of overlapping units
BLEU automated metrics definition?
Similarity between a generated text and one or more reference translations
Used to evaluate the quality of text that has been machine-translated from one natural language to another
BERTScore automated metrics definition?
Compute contextualized embeddings for the input texts, and then calculates the cosine similarity between them
It relies on semantic similarity rather than relying on exact lexical matches
ROUGE Vs BLEU Vs BERTScore Vs Perplexity
Compare N-gram matches
Vs
Evaluate Quality(Prcesion and penalizes)
Vs
Semantic similarity(Compare embeddings)
Vs
How confident the model to predict next token(lower is better)
What is automated metrics used for?
Assess the performace of a FM in text summarization, machine translation, and open-ended text generation
What is negative prompting?
Negative prompting is used to guide the model away from producing certain types of content or exhibiting specific behaviors
What are the key elements in a prompt?
Intructions, Context, Input data and desired output
What is used to influence the model response?
Inference parameters
Randomness and diversity parameters?
Temperature - A higher temperature makes the output more diverse and unpredictable, and a lower temperature makes it more focused and predictable
Top P - With a low top p setting, like 0.250, the model will only consider words that make up the top 25 percent of the total probability distribution. Higher P means more diverse
Top K - Set to 50, the model will only consider the 50 most likely words for the next word in the sequence
Length and stop sequence in inference parameter?
Maximum length - Used in text summarization and translation
Stop Sequence - When the model encounters a stop sequence during the inference process, it will terminate the generation regardless of the maximum length setting
Inferencing at the edge?
Running Small Language Models on an edge device
Zero-shot, few shots and CoT prompting?
Zero-shot - Present task to generative model w/o and example
Few-shot - Present task to generative model with some examples
Chain of Thought - Divides intricate reasoning tasks into smaller, intermediary steps
Prompt misuse - Poisoning, hijacking, and prompt injection
Poisoning - intentional introduction of malicious or biased data
Hijacking, and prompt injection - influencing the outputs of generative models by embedding specific instructions
Prompt misuse - Exposure
Risk of exposing sensitive or confidential information from its training corpus
Prompt misuse - Prompt leaking
Exposing the prompt or inputs used within the model or data used by the model
Prompt misuse - Jailbreaking
Modifying or circumventing the constraints and safety measures implemented in a generative model or AI assistant to gain unauthorized access or functionality
The most common algorithms used to perform the similarity search are?
k-NN or cosine similarity
Name a few vector databases offered by Amazon?
Amazon Opesearch
pgvector extension in RDS
Amazon Kendra
What is measured by benchmark databases in model evaluation?
Accuracy
Speed and efficiency
Scalability
Is it true - Creating a benchmark dataset is a manual process?
True. It is a set of questions and answers provided by the SME. Model’s response to the same questions is compared with benchmark datasets answers and model performance is scored
What type of fine tuning is this - retraining the model on a new dataset that consists of prompts followed by the desired outputs?
Instruction tuning
What type of fine tuning is this - model is refined through a reinforcement learning process, where a reward model built from human feedback guides the model?
Reinforcement learning from human feedback (RLHF)
What type of fine tuning is this - the approach involves extending the training phase of a pre-trained model by continuously feeding it new and emerging data?
Continuous pretraining
What are the steps in fine tuning?
Data curation
Labeling
Governance and compliance
Representativeness and bias checking
Feedback integration
What are the two types of ROUGE metrics
ROUGE-N - This metric primarily assesses the fluency of the text and the extent to which it includes key ideas from the reference. Compare N-gram matches between required vs actual output
ROUGE-L - It is good at evaluating the coherence and order of the narrative in the outputs. Compare the longest sequence of words matche between required vs actual output
What does BLEU do?
Measures the precision of N-grams in the machine-generated text that appears in the reference texts and applies a penalty for overly short translations (brevity penalty)
Services that provide infrastructure protection in AWS?
iAM and NACLs
Advantages of AI governance
Managing, optimizing, and scaling the organizational AI initiative
Maintaining responsible and trustworthy AI practices
Establish clear policies, guidelines, and oversight mechanisms
Proximity of data to the compute resources used for training and inference, which AI data governance data management concepts?
Data residency
Model performance metrics and System events, which AI data governance data management concepts?
Data logging
Data visualization and Exploratory data analysis (EDA), which AI data governance data management concepts?
Data analysis
Performance metrics used for monitoring under AI governance
Accuracy
Precision
Recall
F1-score
Latency
Fine tuned models vs Self trained models
Fine tuning a model using your data vs training a model from scratch using your data
What is source citation in GenerativeAI?
It refers to the act of properly attributing and acknowledging the sources of the data used to train the model.
Datasets
Databases
Other sources
What is Documenting data origins in GenerativeAI?
It provides detailed information about the provenance, or the place of origin of the data used to train the model.
Details about the data collection process
The methods used to curate and clean the data
Any preprocessing or transformations applied to the data
What is the technique used to track the history of data, including its origin, transformation, and movement through different systems?
Data lineage
What is the systematic organization and documentation of the datasets, models, and other resources used in the development of a generative AI system?
Cataloging
What is called the standardized format for documenting the key details about an ML model, including its intended use, performance characteristics, and potential limitations?
Model cards
GenerativeAI scoping matrix(From low to high ownership)
Scope 1 : Consumer App (ChatGpt)
Scope 2 : Enterprise App (SaaS like Amazon Q developer)
Scope 3: Pre-trained models (Amazon Bedrock)
Scope 4: Fine-tuned models (Amazon Bedrock customized or SageMaker Jumpstart)
Scope 5: Self trained models (SageMaker )
What is data engineering life cycle
Automation and access control - AWS Glue
Data collection - Kinesis, DMS, Glue
Data prep and cleaning - EMR or Glue
Data quality check - Glue data brew or Glue data quality check
Data visualization and analysis - Quicksight or Neptune
IaC deployment - CloudFormation
Monitoring and Debugging - CloudWatch
4 ways to build your AI solutions by chosing FM?
Reuse
Adapt
Customize
Start from scratch
Reuse and Adapt only in which AWS GenAI tool?
Amazon Q
Amazon Q pricing
Subscription based - Lite and Pro + Data storage for client documents
Amazon Q Admin controls?
Same as Guardrails
Reuse, Adapt and Customize only in which AWS AL/ML service?
Amazon Bedrock
Reuse, Adapt, Customize and Start from scratch
Amazon SageMaker
Foundation Models comparison
Amazon Titan Text Express, LLAMA 2, Claude, stability.ai
Claude can take maximum tokens - 200K
stability.ai is for image
Content creation is by Titan
Text generation and customer service by LLAMA 2
Analysis and Forecasting by Claude
Image creation by stability.ai
What is PartyRock
Its a playground on Amazon Bedrock to build GenAI apps
You can access without having AWS account
Which model used in classification?
KNN
Custom classfication and custom entity recognition in which AWS AI managed service?
Amazon Comprehend
Redaction is used in which AWS AI managed service?
Amazon Transcribe to remove PII information
Custome vocabularies and custome language models are used in which AWS AI managed service?
Amazon Transcribe to transcribe technical terms and jargons and context
Lexicons, SSML are used in which AWS AI managed service?
Amazon Polly to read specific type of text and add break, whisper etc
What are SLOTs use in which AWS AI managed service?
Amazon Lex to provide input parameters
Amazon Lex can integrate with what AWS services?
Comprehend, Connect, Lambda function & Kendra