AI Prac Terms Flashcards
is a centralized logging service that monitors AWS resources and stores application logs and performance metrics. You can use CloudWatch to monitor and observe resources
Amazon CloudWatch
Data Preparation, Transformation and feature engineering Tool
SageMaker Data Wrangler
These are a feature of SageMaker that you can use to record information about ML models. It includes information such as training details, evaluation metrics, and model performance.
Amazon SageMaker Model Cards
build ML models without needing to write any code. It does not have any models that can perform content moderation of creative content types.
SageMaker Canvas
is a service that uses a human workforce to create accurate labels for data that you can use to train models. does not store information about model training and performance for audit purposes.
SageMaker Ground Truth
This monitors the quality of SageMaker machine learning models in production. You can set up continuous monitoring with a real-time endpoint (or a batch transform job that runs regularly), or on-schedule monitoring for asynchronous batch transform jobs.
SageMaker Model Monitor
is a feature that you can use when you create generative AI applications. They can automatically call Amazon Bedrock APIs and can enhance foundation model (FM) performance. They do not store information about model training and performance for audit purposes. They do Task Coordination, Leverage RAG.
Agents for Amazon Bedrock
is a feature that helps manage generative AI applications. They filter out unwanted topics or content and add safeguard to the model. They do not store information about model training and performance for audit purposes.
Guardrails for Amazon Bedrock
offers a suite of integrated development environments (IDEs), including JupyterLab, RStudio, and Visual Studio Code - Open Source (Code-OSS). You can use it to build content moderation models that can handle creative content types. However, this solution requires additional operational overhead.
SageMaker Studio
is a fully managed AI service for image and video analysis. You can use it to identify inappropriate content in images, including drawings, paintings, and animations. it is designed specifically for performing content moderation of the creative content types. Additionally, you can access it directly through an API. Therefore, it requires the least operational overhead.
Amazon Rekognition
indicates that the model is not making erroneous assumptions about the training data. indicates that the model is not paying attention to noise in the training data. This is an ideal outcome for model training and would not result in model overfitting / underfitting.
Low bias & Low variance
indicates that the model is not making erroneous assumptions about the training data. indicates that the model is paying attention to noise in the training data and is overfitting. When a model performs well on training data but fails to generalize to new data
Low bias & High variance
indicates that the model is making erroneous assumptions about the training data. indicates that the model is not paying attention to noise in the training data, which will lead to underfitting
High bias & Low variance
is an audit resource that provides on-demand access to security and compliance documentation for the AWS Cloud.
AWS Artifact
provides resources and recommendations for cost optimization, security, and resilience. It evaluates your AWS environment, compares environment settings with best practices, and recommends actions to remediate any deviation from best practices.
Trusted Advisor
is a service that tracks user activity and API usage on AWS. You can use it for audit purposes to record actions taken by users, roles, and services in your AWS account.
CloudTrail
is a centralized logging service that monitors AWS resources and stores application logs and performance metrics. You can use it to monitor and observe resources
CloudWatch
uses ML to discover, monitor, and protect sensitive data that is stored in Amazon S3. You can use it to identify and protect PII. You can use it to comply with data governance and privacy regulations
Macie
provides an overview of your AWS resource configurations. You can use it to identify how resources were configured in the past. it can identify settings that do not meet compliance standards, such as if an S3 bucket is publicly accessible.
AWS Config
is a fully managed, native JSON document database. You can use it to operate critical document workloads at scale without the need to manage infrastructure. it supports vector search. You can use vector search to store, index, and search millions of vectors with millisecond response times. it can perform real-time similarity queries with low latency.
Amazon DocumentDB
is a fully managed service that you can use to deploy, scale, and operate it on AWS. You can use it vector database capabilities for many purposes. For example, you can implement semantic search, retrieval augmented generation (RAG) with large language models (LLMs), recommendation engines, and multimedia searches. It supports storing vector embeddings for similarity search capabilities with low latency. It can also scale to store millions of embeddings and can support high query throughput.
OpenSearch Service
is a feature of SageMaker that helps you explain how a model makes predictions and whether datasets or models reflect bias. It also includes a library to evaluate FM performance. The foundation model evaluation (FMEval) library includes tools to compare FM quality and responsibility metrics, including bias and toxicity scores. FMEval can use built-in test datasets, or you can provide a test dataset that is specific to your use case.
SageMaker Clarify
is a hub that consists of hundreds of open source pre-trained models for a wide range of problem types. However, a company cannot insert its models into it.
SageMaker JumpStart
is a fully managed catalog for ML models. You can use it to manage model versions, associate metadata with models, and manage model approval status. You can use SageMaker Canvas to push built models to it. SageMaker Studio users can then access the same thing and the models in the registry. This solution requires the least operational overhead because the company needs only to register the models to implement the workflow.
SageMaker Model Registry