AI Practice Test #4 Flashcards

1
Q

Asynchronous inference

A

Asynchronous inference is the most suitable choice for this scenario. It allows the company to process smaller payloads without requiring real-time responses by queuing the requests and handling them in the background. This method is cost-effective and efficient when some delay is acceptable, as it frees up resources and optimizes compute usage. Asynchronous inference is ideal for scenarios where the payload size is less than 1 GB and immediate results are not critical.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Batch inference

A

Batch inference is generally used for processing large datasets all at once. While it does not require immediate responses, it is typically more efficient for handling larger payloads (several gigabytes or more). For smaller payloads of less than 1 GB, batch inference might be overkill and less cost-efficient compared to asynchronous inference.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Real-time inference

A

Real-time inference is optimized for scenarios where low latency is essential, and responses are needed immediately. It is not suitable for cases where the system can afford to wait for responses, as it might lead to higher costs and resource consumption without providing any additional benefit for this particular use case.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Serverless inference

A

Serverless inference is a good choice for workloads with unpredictable traffic or sporadic requests, as it scales automatically based on demand. However, it may not be as cost-effective for scenarios where workloads are predictable, and some waiting time is acceptable. Asynchronous inference provides a more targeted solution for handling delayed responses at a lower cost.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are the key constituents of a good prompting technique in this context?

A

Instructions, Context, Input data, Output Indicator

The following are the constituents of a good prompting technique:

Instructions – a task for the model to do (description, how the model should perform)

Context – external information to guide the model

Input data – the input for which you want a response

Output Indicator – the output type or format

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Hyperparameters

A

Hyperparameters are values that can be adjusted for model customization to control the training process and, consequently, the output custom model. In other words, hyperparameters are external configurations set before the training process begins. They control the training process and the structure of the model but are not adjusted by the training algorithm itself. Examples include the learning rate, the number of layers in a neural network, etc.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Model parameters

A

Model parameters are values that define a model and its behavior in interpreting input and generating responses. Model parameters are controlled and updated by providers. You can also update model parameters to create a new model through the process of model customization. In other words, Model parameters are the internal variables of the model that are learned and adjusted during the training process. These parameters directly influence the output of the model for a given input. Examples include the weights and biases in a neural network.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What type of data should be included in the few-shots examples?

A

The data should include user-input along with the correct user intent, providing examples of user queries and the corresponding intent

This is the correct answer because few-shots prompting involves providing the model with examples that include both the user-input and the correct user intent. These examples help the model understand and learn how to map various user queries to their appropriate intents. By repeatedly seeing this pairing, the model can generalize from these examples and improve its ability to recognize user intent in new, unseen queries.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Retrieval-Augmented Generation

A

Utilize a Retrieval-Augmented Generation (RAG) system by indexing all product catalog PDFs and configuring the LLM chatbot to reference this system for answering queries

Using a RAG approach is the least costly and most efficient solution for providing up-to-date and relevant responses. In this approach, you convert all product catalog PDFs into a searchable knowledge base. When a customer query comes in, the RAG framework first retrieves the most relevant pieces of information from this knowledge base and then uses an LLM to generate a coherent response based on the retrieved context. This method does not require re-training the model or modifying every incoming query with large datasets, making it significantly more cost-effective. It ensures that the chatbot always has access to the most recent information without needing expensive updates or processing every time.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Stable Diffusion

A

Stable Diffusion is a generative artificial intelligence (generative AI) model that produces unique photorealistic images from text and image prompts.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Llama

A

Llama is a series of large language models trained on publicly available data. They are built on the transformer architecture, enabling them to handle input sequences of any length and produce output sequences of varying lengths. A notable feature of Llama models is their capacity to generate coherent and contextually appropriate text.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Jurassic

A

Jurassic family of models from AI21 Labs supported use cases such as question answering, summarization, draft generation, advanced information extraction, and ideation for tasks requiring intricate reasoning and logic.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Claude

A

Claude is Anthropic’s frontier, state-of-the-art large language model that offers important features for enterprises like advanced reasoning, vision analysis, code generation, and multilingual processing.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Amazon Comprehend

A

Amazon Comprehend is a natural language processing (NLP) service that uses machine learning to uncover insights and relationships in text. It is specifically designed for tasks such as sentiment analysis, entity recognition, key phrase extraction, and language detection. For the scenario of analyzing customer reviews, Amazon Comprehend can directly determine the overall sentiment of a text (positive, negative, neutral, or mixed), making it the ideal service for this purpose. By using Amazon Comprehend, e-commerce platforms can effectively analyze customer feedback, understand customer satisfaction levels, and identify common themes or concerns.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Amazon Bedrock

A

Amazon Bedrock is an AI service that provides access to foundation models (large language models, including those for NLP tasks) via an API. While Amazon Bedrock is not specifically an NLP service like Amazon Comprehend, it can be used to fine-tune pre-trained foundation models for various tasks, including sentiment analysis. With the proper configuration and fine-tuning, Bedrock can analyze text data to determine sentiment, making it a versatile option for advanced users who may need more customizable solutions than Amazon Comprehend.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Amazon Rekognition

A

Amazon Rekognition is a service designed for analyzing images and videos, not text. It can identify objects, people, text within images, and even detect inappropriate content in images and videos. However, it does not provide any capabilities for natural language processing or sentiment analysis, making it unsuitable for analyzing written customer reviews.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Amazon Textract

A

Amazon Textract is an OCR (Optical Character Recognition) service that extracts printed or handwritten text from scanned documents, PDFs, and images. It is useful for digitizing text but does not offer any features for analyzing or interpreting the sentiment of the extracted text. Since Textract focuses on text extraction rather than understanding or analyzing the content, it is not suitable for sentiment analysis tasks.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Amazon Personalize

A

Amazon Personalize is a service that provides personalized recommendations, search, and ranking for websites and applications based on user behavior and preferences. While it can help improve customer experience by suggesting products or content based on historical data, it does not offer natural language processing or sentiment analysis capabilities. Thus, it is not the correct choice for analyzing written customer reviews to determine sentiment.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Model invocation logging

A

The company should enable model invocation logging, which allows for detailed logging of all requests and responses during model invocations in Amazon Bedrock

You can use model invocation logging to collect invocation logs, model input data, and model output data for all invocations in your AWS account used in Amazon Bedrock. With invocation logging, you can collect the full request data, response data, and metadata associated with all calls performed in your account. Logging can be configured to provide the destination resources where the log data will be published. Supported destinations include Amazon CloudWatch Logs and Amazon Simple Storage Service (Amazon S3). Only destinations from the same account and region are supported. Model invocation logging is disabled by default.

This is the correct option because enabling invocation logging on Amazon Bedrock allows the company to capture detailed logs of all model requests and responses, including input data, output predictions, and any errors that occur during model execution. This method provides comprehensive monitoring capabilities, enabling the company to effectively track, audit, and troubleshoot model performance and usage.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

AWS CloudTrail

A

While AWS CloudTrail is useful for tracking API calls and monitoring who accessed which AWS resources, it does not capture the actual input and output data involved in model invocations. CloudTrail logs are primarily intended for auditing access and managing security rather than monitoring detailed data flow or model performance on Amazon Bedrock.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Amazon EventBridge

A

Amazon EventBridge is designed to react to changes and events across AWS resources and trigger workflows or automate responses. Although it can track when a model invocation occurs, it does not provide detailed logging of the input and output data associated with these invocations, limiting its usefulness for comprehensive monitoring purposes.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

AWS Config

A

AWS Config is specifically designed for monitoring and managing AWS resource configurations and compliance, not for tracking or logging the input and output data of machine learning models on Amazon Bedrock. AWS Config focuses on configuration management and does not provide the level of detail required to monitor data traffic or model performance in machine learning applications.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Generative Adversarial Network (GAN)

A

The company should use a Generative Adversarial Network (GAN) for creating realistic synthetic data while preserving the statistical properties of the original data

This is the correct answer because GANs are specifically designed for generating synthetic data that is statistically similar to real data. They consist of two neural networks—a generator and a discriminator—that work against each other to create highly realistic synthetic data. GANs have been successfully used in various domains, including image generation, text synthesis, and more, to produce data that retains the underlying patterns and structures of the original dataset, making them highly suitable for this purpose.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Support Vector Machines (SVMs)

A

SVMs are used for classification and regression, where the algorithm finds the optimal hyperplane that best separates different classes in the data. SVMs do not generate new data or create synthetic datasets, so they are not suitable for a task that requires generating synthetic data based on existing datasets.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

Convolutional Neural Network (CNN)

A

CNNs are designed for tasks such as image and video recognition, object detection, and similar applications involving grid-like data (such as pixels in an image). While CNNs are excellent at feature extraction and classification in images, they are not suitable for generating synthetic data, especially for non-visual data types.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

WaveNet

A

WaveNet is tailored for audio data generation, specifically for tasks such as speech synthesis and audio signal processing. While it is powerful within its specific domain, it is not designed for generating synthetic data outside of audio, making it an unsuitable choice for general-purpose data generation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

Exploratory Data Analysis (EDA)

A

The company is in the Exploratory Data Analysis (EDA) phase, which involves examining the data through statistical summaries and visualizations to identify patterns, detect anomalies, and form hypotheses. This phase is crucial for understanding the dataset’s structure and characteristics, making it the most appropriate description of the current activities. Tasks like calculating statistics and visualizing data are fundamental to EDA, helping to uncover patterns, detect outliers, and gain insights into the data before any modeling is done. EDA serves as the foundation for building predictive models by providing a deep understanding of the data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

Data Preparation

A

data preparation involves cleaning and preprocessing the data to make it suitable for analysis or modeling. This may include handling missing values, removing duplicates, or transforming variables, but it does not typically involve calculating statistics and visualizing data. While data preparation is an important step, it does not encompass the exploratory analysis activities described in the question.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

Data Augmentation

A

Data augmentation is a technique used primarily in machine learning to artificially increase the size and variability of the training dataset by creating modified versions of the existing data, such as flipping images or adding noise. It is not related to the tasks of calculating statistics or visualizing data, which are part of EDA.

30
Q

Model Evaluation

A

Model evaluation refers to assessing the performance of a machine learning model using specific metrics such as accuracy, precision, recall, or F1 score. Model evaluation does not involve exploratory tasks like calculating statistics or visualizing data; instead, it focuses on validating the effectiveness of a trained model. Therefore, this phase does not align with the company’s current activities.

31
Q

Large language models (LLMs)

A

Large language models (LLM) are very large deep learning models that are pre-trained on vast amounts of data. The underlying transformer is a set of neural networks that consist of an encoder and a decoder with self-attention capabilities. The encoder and decoder extract meanings from a sequence of text and understand the relationships between words and phrases in it.

Large language models (LLMs) are one class of Foundation Models. For example, OpenAI’s generative pre-trained transformer (GPT) models are LLMs. LLMs are specifically focused on language-based tasks such as such as summarization, text generation, classification, open-ended conversation, and information extraction.

AWS recommends AWS Bedrock and Amazon SageMaker JumpStart as the best-fit services for developing LLM based solutions.

32
Q

Amazon Bedrock

A

Amazon Bedrock is the easiest way to build and scale generative AI applications with foundation models. Amazon Bedrock is a fully managed service that makes foundation models from Amazon and leading AI startups available through an API, so you can choose from various FMs to find the model that’s best suited for your use case. With Bedrock, you can speed up developing and deploying scalable, reliable, and secure generative AI applications without managing infrastructure.

33
Q

Amazon SageMaker JumpStart

A

Amazon SageMaker JumpStart is a machine learning hub with foundation models, built-in algorithms, and prebuilt ML solutions that you can deploy with just a few clicks. With SageMaker JumpStart, you can access pre-trained models, including foundation models, to perform tasks like article summarization and image generation. Pretrained models are fully customizable for your use case with your data, and you can easily deploy them into production with the user interface or SDK.

34
Q

Amazon Q

A

Amazon Q is a generative AI–powered assistant for accelerating software development and leveraging companies’ internal data. Amazon Q generates code, tests, and debugs. It has multistep planning and reasoning capabilities that can transform and implement new code generated from developer requests.

35
Q

AWS Trainium

A

AWS Trainium is the machine learning (ML) chip that AWS purpose-built for deep learning (DL) training of 100B+ parameter models. Each Amazon Elastic Compute Cloud (Amazon EC2) Trn1 instance deploys up to 16 Trainium accelerators to deliver a high-performance, low-cost solution for DL training in the cloud.

36
Q

AWS Inferentia

A

AWS Inferentia is an ML chip purpose-built by AWS to deliver high-performance inference at a low cost. AWS Inferentia accelerators are designed by AWS to deliver high performance at the lowest cost in Amazon EC2 for your deep learning (DL) and generative AI inference applications.

37
Q

Foundation Models (FMs)

A

FMs use self-supervised learning to create labels from input data, however, fine-tuning an FM is a supervised learning process

Foundation models use self-supervised learning to create labels from input data. This means no one has instructed or trained the model with labeled training data sets.

38
Q

fine-tuning

A

fine-tuning an FM is a supervised learning process

Fine-tuning a pre-trained foundation model is an affordable way to take advantage of their broad capabilities while customizing a model on your own small, corpus. Fine-tuning involves further training a pre-trained language model on a specific task or domain-specific dataset, allowing it to address business requirements. Fine-tuning is a customization method that does change the weights of your model.

Fine-tuning an FM is a supervised learning process.

39
Q

supervised learning

A

In supervised learning, you train the model with a set of input data and a corresponding set of paired labeled output data.

40
Q

Unsupervised learning

A

Unsupervised machine learning is when you give the algorithm input data without any labeled output data. Then, on its own, the algorithm identifies patterns and relationships in and between the data.

41
Q

Self-supervised learning

A

Self-supervised learning is a machine learning approach that applies unsupervised learning methods to tasks usually requiring supervised learning. Instead of using labeled datasets for guidance, self-supervised models create implicit labels from unstructured data.

42
Q

Shapley values

A

Shapley values provide a local explanation by quantifying the contribution of each feature to the prediction for a specific instance

Use Shapley values to explain individual predictions

Shapley values are a local interpretability method that explains individual predictions by assigning each feature a contribution score based on its marginal effect on the prediction. This method is useful for understanding the impact of each feature on a specific instance’s prediction.

43
Q

Partial Dependence Plots (PDP)

A

PDP provides a global explanation by showing the marginal effect of a feature on the model’s predictions across the dataset.

PDP to understand the model’s behavior at a dataset level

Partial Dependence Plots (PDP), on the other hand, provide a global view of the model’s behavior by illustrating how the predicted outcome changes as a single feature is varied across its range, holding all other features constant. PDPs help understand the overall relationship between a feature and the model output across the entire dataset.

44
Q

Amazon Q in QuickSight

A

With Amazon Q in QuickSight, customers get a generative BI assistant that allows business analysts to use natural language to build BI dashboards in minutes and easily create visualizations and complex calculations. These dashboard-authoring capabilities empower business analysts to swiftly build, uncover, and share valuable insights using natural language prompts. You can simplify data understanding for business users through a context-aware Q&A experience, executive summaries, and customizable data stories — all designed to use insights to inform and drive decisions.

45
Q

Amazon Q Developer

A

Amazon Q Developer assists developers and IT professionals with all their tasks—from coding, testing, and upgrading applications, to diagnosing errors, performing security scanning and fixes, and optimizing AWS resources.

46
Q

Amazon Q Business

A

Amazon Q Business is a fully managed, generative-AI-powered assistant that you can configure to answer questions, provide summaries, generate content, and complete tasks based on your enterprise data. It allows end users to receive immediate, permissions-aware responses from enterprise data sources with citations, for use cases such as IT, HR, and benefits help desks.

47
Q

Amazon Q in Connect

A

Amazon Connect is the contact center service from AWS. Amazon Q helps customer service agents provide better customer service. Amazon Q in Connect enriches real-time customer conversations with the relevant company content. It recommends what to say or what actions an agent should take to assist customers in a better way.

48
Q

Amazon Comprehend

A

Amazon Comprehend is a natural language processing (NLP) service that uses machine learning to find insights and relationships in text, no machine learning experience is required. Amazon Comprehend uses machine learning to help you uncover the insights and relationships in your unstructured data.

By using Amazon Comprehend on the converted text data from Amazon Transcribe, you can perform sentiment analysis or extract entities and key phrases.

49
Q

sentiment analysis

A

Sentiment analysis is the process of analyzing digital text to determine if the emotional tone of the message is positive, negative, or neutral.

50
Q

Confusion matrix

A

Confusion matrix is a tool specifically designed to evaluate the performance of classification models by displaying the number of true positives, true negatives, false positives, and false negatives. This matrix provides a detailed breakdown of the model’s performance across all classes, making it the most suitable choice for evaluating a classification model’s accuracy and identifying potential areas for improvement. It provides a comprehensive overview of the model’s performance by detailing how many instances were correctly or incorrectly classified in each category. This enables the company to understand where the model is performing well and where it may need adjustments, such as improving the classification of specific material types.

51
Q

Root Mean Squared Error (RMSE)

A

Root Mean Squared Error (RMSE) is a metric commonly used to measure the average error in regression models by calculating the square root of the average squared differences between predicted and actual values. However, RMSE is not suitable for classification tasks, as it is designed to measure continuous outcomes, not discrete class predictions.

52
Q

Mean Absolute Error (MAE)

A

Mean Absolute Error (MAE) measures the average magnitude of errors in a set of predictions without considering their direction. MAE is typically used in regression tasks to quantify the accuracy of a continuous variable’s predictions, not for classification tasks where the outputs are categorical rather than continuous.

53
Q

Correlation matrix

A

Correlation matrix measures the statistical correlation between different variables or features in a dataset, typically used to understand the relationships between continuous variables. A correlation matrix is not designed to evaluate the performance of a classification model, as it does not provide any insight into the accuracy or errors of categorical predictions.

54
Q

Transformer models

A

Transformer models use a self-attention mechanism and implement contextual embeddings

Transformer models are a type of neural network architecture designed to handle sequential data, such as language, in an efficient and scalable way. They rely on a mechanism called self-attention to process input data, allowing them to understand and generate language effectively. Self-attention allows the model to weigh the importance of different words in a sentence when encoding a particular word. This helps the model capture relationships and dependencies between words, regardless of their position in the sequence.

Transformer models use self-attention to weigh the importance of different words in a sentence, allowing them to capture complex dependencies. Positional encodings provide information about word order, and the encoder-decoder architecture enables effective processing and generation of sequences. This makes transformers highly effective for tasks like language translation, text generation, and more.

55
Q

Generative Adversarial Networks (GANs)

A

Generative Adversarial Networks (GANs) work by training two neural networks in a competitive manner. The first network, known as the generator, generates fake data samples by adding random noise. The second network, called the discriminator, tries to distinguish between real data and the fake data produced by the generator.

56
Q

Variational autoencoders (VAEs)

A

Variational autoencoders (VAEs) learn a compact representation of data called latent space. You can think of it as a unique code representing the data based on all its attributes. VAEs use two neural networks—the encoder and the decoder. The encoder neural network maps the input data to a mean and variance for each dimension of the latent space. The decoder neural network takes this sampled point from the latent space and reconstructs it back into data that resembles the original input.

57
Q

Diffusion models

A

Diffusion models work by first corrupting data with noise through a forward diffusion process and then learning to reverse this process to denoise the data. They use neural networks to predict and remove the noise step by step, ultimately generating new, structured data from random noise.

58
Q

reinforcement learning (RL)

A

The company should leverage reinforcement learning (RL), where rewards are generated from positive customer feedback to train the chatbot in optimizing its responses

Reinforcement learning is the most suitable approach for self-improvement in this context. By leveraging RL, the chatbot can learn from customer interactions in real-time. Positive customer feedback serves as a reward signal that guides the chatbot to improve its responses over time. The chatbot adapts its behavior based on rewards or penalties, refining its conversational skills through continuous feedback loops. This dynamic learning process is effective for environments where responses need to be optimized based on direct user interaction and satisfaction.

59
Q

epochs

A

an epoch is a complete pass of a training dataset through a learning algorithm, also known as a training cycle.

The company should increase the number of epochs, which involves training the model for more iterations over the dataset

Increasing the number of epochs allows the model to learn from the training data for a longer period, potentially capturing more complex patterns and relationships, which can improve accuracy. Multiple epochs are run until the accuracy of the model reaches an acceptable level, or when the error rate drops below an acceptable level.

60
Q

MLflow with Amazon SageMaker

A

Manage machine learning experiments

Machine learning is an iterative process that requires experimenting with various combinations of data, algorithms, and parameters while observing their impact on model accuracy. The iterative nature of ML experimentation results in numerous model training runs and versions, making it challenging to track the best-performing models and their configurations.

Use MLflow with Amazon SageMaker to track, organize, view, analyze, and compare iterative ML experimentation to gain comparative insights and register and deploy your best-performing models.

61
Q

Large Language Model (LLM)

A

Large language models (LLMs) are a class of Foundation Models (FMs). For example, OpenAI’s generative pre-trained transformer (GPT) models are LLMs. LLMs are specifically focused on language-based tasks such as such as summarization, text generation, classification, open-ended conversation, and information extraction.

62
Q

Retrieval-Augmented Generation (RAG)

A

Retrieval-Augmented Generation (RAG) is the process of optimizing the output of a large language model, so it references an authoritative knowledge base outside of its training data sources before generating a response. Large Language Models (LLMs) are trained on vast volumes of data and use billions of parameters to generate original output for tasks like answering questions, translating languages, and completing sentences. RAG extends the already powerful capabilities of LLMs to specific domains or an organization’s internal knowledge base, all without the need to retrain the model. It is a cost-effective approach to improving LLM output so it remains relevant, accurate, and useful in various contexts.

Depending on the configuration, Amazon Q Business web application workflow can use LLM/RAG or both.

63
Q

Diffusion Model

A

Diffusion models create new data by iteratively making controlled random changes to an initial data sample. They start with the original data and add subtle changes (noise), progressively making it less similar to the original. This noise is carefully controlled to ensure the generated data remains coherent and realistic. After adding noise over several iterations, the diffusion model reverses the process. Reverse denoising gradually removes the noise to produce a new data sample that resembles the original.

64
Q

Generative adversarial network (GAN)

A

GANs work by training two neural networks in a competitive manner. The first network, known as the generator, generates fake data samples by adding random noise. The second network, called the discriminator, tries to distinguish between real data and the fake data produced by the generator. During training, the generator continually improves its ability to create realistic data while the discriminator becomes better at telling real from fake. This adversarial process continues until the generator produces data that is so convincing that the discriminator can’t differentiate it from real data.

65
Q

Variational autoencoders (VAE)

A

VAEs use two neural networks—the encoder and the decoder. The encoder neural network maps the input data to a mean and variance for each dimension of the latent space. It generates a random sample from a Gaussian (normal) distribution. This sample is a point in the latent space and represents a compressed, simplified version of the input data. The decoder neural network takes this sampled point from the latent space and reconstructs it back into data that resembles the original input.

66
Q

Amazon Bedrock Guardrails

A

The company should instruct the model to stick to the prompt by adding explicit instructions to ignore any unrelated or potentially malicious content

This is the correct approach because providing explicit instructions within the prompt helps guide the model’s behavior, reducing the likelihood of generating inappropriate or unsafe content. By clarifying what the model should focus on and what it should ignore, the company can enforce boundaries that align with its safety standards. This method is straightforward and leverages prompt engineering to mitigate risks effectively.

67
Q

Generative AI powered summarization chatbot

A

Generative AI powered summarization chatbot leverages large language models to generate concise summaries of text. With prompt engineering, the summarization chatbot can be specifically tailored to accurately extract detailed key points, entities, or legal clauses from complex legal documents.

68
Q

WaveNet

A

WaveNet is tailored for audio data generation, specifically for tasks such as speech synthesis and audio signal processing. It does not have the capabilities to analyze legal documents or extract key information, making it an incorrect choice for this task.

69
Q

Domain Adaptation Fine-Tuning

A

The company should use Domain Adaptation Fine-Tuning, which involves fine-tuning the model on domain-specific data to adapt its knowledge to that particular domain

Domain Adaptation Fine-Tuning is an effective approach because it takes a pre-trained Foundation Model and further adjusts its parameters using domain-specific data. This process helps the model learn the nuances, terminology, and context specific to the domain, enhancing its ability to generate accurate and relevant outputs in that field. Fine-tuning allows the model to specialize while retaining the general knowledge acquired during initial training.

70
Q

Continued Pre-Training

A

The company should use Continued Pre-Training, which involves further training the model on a large corpus of domain-specific data, enhancing its ability to understand domain-specific terms, jargon, and context

Continued Pre-Training is another appropriate strategy for making a Foundation Model an expert in a specific domain. By pre-training the model on a large dataset specifically from the target domain, the model can learn the distinct characteristics, language patterns, and specialized knowledge relevant to that domain. This approach effectively builds upon the model’s existing knowledge, enhancing its domain expertise without starting training from scratch.