Other Flashcards
Amazon Comprehend
Amazon Comprehend is a natural language processing (NLP) service that uses machine learning to uncover insights and relationships in text. It is specifically designed for tasks such as sentiment analysis, entity recognition, key phrase extraction, and language detection.
Amazon Kendra
Machine learning tool that searches websites and applications to help users search content
Llama
a series of large language models trained on publicly available data, and built on the transformer architecture. It can generate coherent and contextually appropriate text.
Jurassic
Family of models from AI21 Labs supported use cases such as question answering, summarization, draft generation, advanced information extraction, and ideation for tasks requiring intricate reasoning and logic.
Claude
Anthropic’s frontier, state-of-the-art large language model that offers important features for enterprises like advanced reasoning, vision analysis, code generation, and multilingual processing.
AWS Trainium
2nd generation machine learning (ML) accelerator for deep learning training. Used by Amazon Elastic Compute Cloud (EC2)
AWS Inferentia
accelerators designed to deliver high performance at the lowest cost in Amazon EC2 for your deep learning (DL) and generative AI inference applications.
Amazon SageMaker GroundTruth Plus
provides a fully managed data labeling service that helps deliver high-quality annotations. It uses a combination of human labelers and machine learning
multi-class vs. multi-label classification
Multi-class classification assigns each instance to one of several possible classes
Multi-label classification assigns each instance to one or more classes
Amazon Stagemaker studio
Offers a broad set of fully managed integrated development environments (IDEs) for ML development, including JupyterLab, Code Editor based on Code-OSS (Visual Studio Code – Open Source), and RStudio.
Amazon Sagemaker -
Asynchronous Inference, Batch transformation & real-time
Asynchronous Inference is used for requests with large payload sizes up to 1GB, long processing times, and near real-time latency requirements.
Batch transform - used to get predictions for an entire dataset
Real-time - For persistent, real-time endpoints that make one prediction at a time
Are large language models deterministic or discriminative?
Large Language Models (LLMs) are non-deterministic, which implies that the generated text may be different for every user that uses the same prompt