AI Practice Test #1 (OLD) Flashcards
(282 cards)
Amazon Bedrock
https://aws.amazon.com/bedrock/agents/
https://aws.amazon.com/bedrock/faqs/
https://docs.aws.amazon.com/bedrock/latest/userguide/general-guidelines-for-bedrock-users.html
Agents for Amazon Bedrock
Agents for Amazon Bedrock are fully managed capabilities that make it easier for developers to create generative AI-based applications that can complete complex tasks for a wide range of use cases and deliver up-to-date answers based on proprietary knowledge sources.
Agents are software components or entities designed to autonomously or semi-autonomously perform specific actions or tasks based on predefined rules or algorithms. With Amazon Bedrock, agents are utilized to manage and execute various multi-step tasks related to infrastructure provisioning, application deployment, and operational activities. For example, you can create an agent that helps customers process insurance claims or an agent that helps customers make travel reservations. You don’t have to provision capacity, manage infrastructure, or write custom code. Amazon Bedrock manages prompt engineering, memory, monitoring, encryption, user permissions, and API invocation.
https://docs.aws.amazon.com/bedrock/latest/userguide/agents.html
Knowledge Bases for Amazon Bedrock
With Knowledge Bases for Amazon Bedrock, you can give FMs and agents contextual information from your company’s private data sources for Retrieval Augmented Generation (RAG) to deliver more relevant, accurate, and customized responses. You cannot use Knowledge Bases for Amazon Bedrock for the given use case.
Watermark detection for Amazon Bedrock
The watermark detection mechanism allows you to identify images generated by Amazon Titan Image Generator, a foundation model that allows users to create realistic, studio-quality images in large volumes and at low cost, using natural language prompts. With watermark detection, you can increase transparency around AI-generated content by mitigating harmful content generation and reducing the spread of misinformation. You cannot use watermark detection for the given use case.
Guardrails for Amazon Bedrock
Guardrails for Amazon Bedrock help you implement safeguards for your generative AI applications based on your use cases and responsible AI policies. It helps control the interaction between users and FMs by filtering undesirable and harmful content, redacts personally identifiable information (PII), and enhances content safety and privacy in generative AI applications. You cannot use Guardrails for Amazon Bedrock for the given use case.
Generative AI
Generative AI can automate the creation of new data based on existing patterns, enhancing productivity and innovation
Generative AI in the AWS cloud environment is advantageous because it automates the creation of new data from existing patterns, which can significantly boost productivity and drive innovation. This capability allows businesses to generate new insights, designs, and solutions more efficiently.
via - https://aws.amazon.com/what-is/generative-ai/
Incorrect options:
Generative AI can replace all human roles in software development - Generative AI is not designed to replace all human roles in software development but to assist and enhance human capabilities by automating certain tasks and creating new data based on patterns. So, this option is incorrect.
Generative AI ensures 100% security against all cyber threats - While generative AI can improve security by identifying patterns and anomalies, it does not guarantee 100% security against all cyber threats. Security in the cloud involves a combination of multiple strategies and tools. Therefore, this option is incorrect.
Generative AI can perform all cloud maintenance tasks without any human intervention - Generative AI can assist in cloud maintenance tasks by predicting issues and suggesting solutions, but it cannot perform all maintenance tasks without human oversight and intervention. So, this option is not the right fit.
References:
https://aws.amazon.com/what-is/generative-ai/
https://aws.amazon.com/ai/generative-ai/services/
Prompt Engineering
https://aws.amazon.com/what-is/prompt-engineering/
Negative Prompting
Negative prompting refers to guiding a generative AI model to avoid certain outputs or behaviors when generating content. In the context of AWS generative AI, like those using Amazon Bedrock, negative prompting is used to refine and control the output of models by specifying what should not be included in the generated content.
Few-shot Prompting
In few-shot prompting, you provide a few examples of a task to the model to guide its output.
Chain-of-thought prompting
Chain-of-thought prompting is a technique that breaks down a complex question into smaller, logical parts that mimic a train of thought. This helps the model solve problems in a series of intermediate steps rather than directly answering the question. This enhances its reasoning ability. It involves guiding the model through a step-by-step process to arrive at a solution or generate content, thereby enhancing the quality and coherence of the output.
Zero-shot Prompting
Zero-shot prompting is a technique used in generative AI where the model is asked to perform a task or generate content without having seen any examples of that specific task during training. Instead, the model relies on its general understanding and knowledge to respond.
GPT
Generative Pre-trained Transformer
The company should use GPT (Generative Pre-trained Transformer), to interpret natural language inputs and generating coherent outputs, such as SQL queries, by leveraging its understanding of language patterns and structures
This is the correct option because GPT models are specifically designed to process and generate human-like text based on context and input data. GPT can be fine-tuned to understand specific domain language and generate accurate SQL queries from plain text input. It uses advanced natural language processing (NLP) techniques to parse input text, understand user intent, and generate the appropriate SQL statements, making it highly suitable for the task.
https://aws.amazon.com/what-is/gpt/
GAN
Generative Adversarial Network
A generative adversarial network (GAN) is a deep learning architecture. It trains two neural networks to compete against each other to generate more authentic new data from a given training dataset. For instance, you can generate new images from an existing image database or original music from a database of songs. A GAN is called adversarial because it trains two different networks and pits them against each other. One network generates new data by taking an input data sample and modifying it as much as possible. The other network tries to predict whether the generated data output belongs in the original dataset. In other words, the predicting network determines whether the generated data is fake or real. The system generates newer, improved versions of fake data values until the predicting network can no longer distinguish fake from original.
via - https://aws.amazon.com/what-is/gan/
Amazon Comprehend
Amazon Comprehend is built for analyzing and extracting insights from text, such as identifying sentiment, entities, and key phrases. It does not have the capability to generate SQL queries from natural language input. Therefore, it does not meet the company’s need for text-to-SQL conversion.
ResNet
Residual Neural Network
ResNet is a deep neural network architecture used mainly in computer vision tasks, such as image classification and object detection. It is not capable of handling natural language input or generating text-based outputs like SQL queries, making it irrelevant to the company’s needs.
WaveNet
WaveNet is a deep generative model created by DeepMind to synthesize audio data, particularly for generating realistic-sounding speech. It is not built to handle text input or produce SQL queries, making it completely unsuitable for this task.
Amazon SageMaker Data Wrangler - Use Case
Fix bias by balancing the dataset
When the number of samples in the majority class (bigger) is considerably larger than the number of samples in the minority (smaller) class, the dataset is considered imbalanced. This skew is challenging for ML algorithms and classifiers because the training process tends to be biased towards the majority class. Data Wrangler supports several balancing operators as part of the Balance data transform.
Incorrect options:
Monitor the quality of a model - This option is incorrect because monitoring model quality is a feature of SageMaker Model Monitor, not SageMaker Data Wrangler. SageMaker Model Monitor is designed to track model quality as well as performance in production.
Build ML models with no code - SageMaker Data Wrangler is not designed for building machine learning models without coding. SageMaker Canvas, another tool in the SageMaker suite, specifically targets no-code model building, allowing users to create and deploy models using a visual interface.
Store and share the features used for model development - SageMaker Feature Store is specifically designed to store and share machine learning features. It allows data scientists and engineers to create a centralized, consistent, and standardized set of features that can be easily accessed and reused across different teams and projects, making it the ideal choice for sharing features during model development. SageMaker Data Wrangler is not designed for this use case.
Reference:
https://aws.amazon.com/blogs/machine-learning/balance-your-data-for-machine-learning-with-amazon-sagemaker-data-wrangler/
Amazon SageMaker Data Wrangler
Amazon SageMaker Data Wrangler
You can split a machine learning (ML) dataset into train, test, and validation datasets with Amazon SageMaker Data Wrangler.
Data used for ML is typically split into the following datasets:
Training – Used to train an algorithm or ML model. The model iteratively uses the data and learns to provide the desired result.
Validation – Introduces new data to the trained model. You can use a validation set to periodically measure model performance as it trains and also tune any hyperparameters of the model. However, validation datasets are optional.
Test – Used on the final trained model to assess its performance on unseen data. This helps determine how well the model generalizes.
Data Wrangler is a capability of Amazon SageMaker that helps data scientists and data engineers quickly and easily prepare data for ML applications using a visual interface. It contains over 300 built-in data transformations so you can quickly normalize, transform, and combine features without writing code.
References:
https://aws.amazon.com/blogs/machine-learning/create-train-test-and-validation-splits-on-your-data-for-machine-learning-with-amazon-sagemaker-data-wrangler/
Amazon SageMaker Clarify
SageMaker Clarify is used to evaluate models and explain the model predictions.
Amazon SageMaker Feature Store
Amazon SageMaker Feature Store is a fully managed, purpose-built repository to store, share, and manage features for machine learning (ML) models.
https://aws.amazon.com/sagemaker/feature-store/
Amazon SageMaker Ground Truth
Amazon SageMaker Ground Truth is a data labeling service provided by AWS that enables users to build highly accurate training datasets for machine learning quickly. The service helps automate the data labeling process through a combination of human labeling and machine learning.
https://aws.amazon.com/sagemaker/groundtruth/
Context window
The context window defines how much text (measured in tokens) the AI model can process at one time to generate a coherent output. It determines the limit of input data that the model can use to understand context, maintain conversation history, or generate relevant responses. The context window is measured in tokens (units of text), not characters, making it the key concept for understanding data processing limits in AI models.
via - https://aws.amazon.com/blogs/security/context-window-overflow-breaking-the-barrier/
Character count
Character count measures the number of characters in a piece of text, but AI models typically do not limit their input based on characters alone. Instead, they rely on tokens, which can represent words, subwords, or punctuation marks. The concept that defines how much text can be processed at one time is the context window, which is measured in tokens, not character count.
Tokens
While tokens are the individual units of text that the model processes, the concept that describes the total amount of text the model can handle at one time is the context window, not tokens themselves. Tokens are components within the context window, and the model’s capacity is defined by how many tokens can fit within this window, rather than just the tokens themselves.