Computer Vision (15-20%) Flashcards
Concepts related to Azure Computer Vision resources (33 cards)
Use this feature for general, unstructured documents with smaller amount of text, or images that contain text.
Azure AI Vision - OCR
Use this service to read small to large volumes of text from images and PDF documents.
Azure AI Document Intelligence
Which service do you use to read text from street signs, handwritten notes, and store signs?
OCR
Which API would be best for this scenario? You need to read a large number of files with high accuracy. The text is short sections of handwritten text, some in English and some of it is in multiple languages.
Image Analysis service OCR feature
What levels of division are the OCR results returned?
Results contain blocks, words and lines, as well as bounding boxes for each word and line.
You’ve scanned a letter into PDF format and need to extract the text it contains. What should you do?
The Document Intelligence API can be used to process PDF formatted files.
You want to use the Azure AI Vision Analyze Image function to generate an appropriate caption for an image. Which visual feature should you specify?
To generate a caption, include the Description visual feature in your analysis.
What is the purpose of the Azure AI Vision service?
The Azure AI Vision service is designed to help you extract information from images through various functionalities.
In AI Vision Service what features are available in the VisualFeatures enum?
VisualFeatures.TAGS: Identifies tags about the image, including objects, scenery, setting, and actions
VisualFeatures.OBJECTS: Returns the bounding box for each detected object
VisualFeatures.CAPTION: Generates a caption of the image in natural language
VisualFeatures.DENSE_CAPTIONS: Generates more detailed captions for the objects detected
VisualFeatures.PEOPLE: Returns the bounding box for detected people
VisualFeatures.SMART_CROPS: Returns the bounding box of the specified aspect ratio for the area of interest
VisualFeatures.READ: Extracts readable text
What functionality exists in Azure Video Indexer?
Facial recognition - detecting the presence of individual people in the image. This requires Limited Access approval.
Optical character recognition - reading text in the video.
Speech transcription - creating a text transcript of spoken dialog in the video.
Topics - identification of key topics discussed in the video.
Sentiment - analysis of how positive or negative segments within the video are.
Labels - label tags that identify key objects or themes throughout the video.
Content moderation - detection of adult or violent themes in the video.
Scene segmentation - a breakdown of the video into its constituent scenes.
What extensions can be made to Azure Video Indexer for custom insights?
People. Add images of the faces of people you want to recognize in videos, and train a model. Video Indexer will then recognize these people in all of your videos.
Note
This only works after Limited Access approval, adhering to our Responsible AI standard.
Language. If your organization uses specific terminology that may not be in common usage, you can train a custom model to detect and transcribe it.
Brands. You can train a model to recognize specific names as brands, for example to identify products, projects, or companies that are relevant to your business.
You want Azure Video Indexer to analyze a video. What must you do first?
Upload the video to Azure Video Indexer and index it.
You want Azure Video Indexer to recognize brands in videos recorded from conference calls. What should you do?
Edit the Brands model to show brands suggested by Bing, and add any new brands you want to detect.
What resources need to be provisiioned to use the AI Custom Vision Service?
A training resource (used to train your models). This can be:
An Azure AI Services resource.
An Azure AI Custom Vision (Training) resource.
A prediction resource, used by client applications to get predictions from your model. This can be:
An Azure AI Services resource.
An Azure AI Custom Vision (Prediction) resource.
Explain multiclass classification
there are multiple classes in the image dataset, but each image can belong to only one class
Explain multilabel classification
an image might be associated with multiple labels
What steps are performed in the Azure AI Custom Vision portal?
Create an image classification project for your model and associate it with a training resource.
Upload images, assigning class label tags to them.
Review and edit tagged images.
Train and evaluate a classification model.
Test a trained model.
Publish a trained model to a prediction resource.
You want to train a model that can categorize an image as “cat” or “dog” based on its subject. What kind of Azure AI Custom Vision project should you create?
To train a model that classifies an image using a single tag, use an Image classification (multiclass) project.
Which of the following kinds of Azure resource can you use to host a trained Azure AI Custom Vision model?
You can publish a trained Azure AI Custom Vision model to either an Azure AI Custom Vision (Prediction) resource, or an Azure AI Services resource.
What features are available in the Face service within Azure AI Vision?
Face detection (with bounding box).
Comprehensive facial feature analysis (including head pose, presence of spectacles, blur, facial landmarks, occlusion and others).
Face comparison and verification.
Facial recognition.
Facial landmark location
Facial liveness - liveness can be used to determine if the input video is a real stream or a fake
What method do you take to detect and analyze faces using the Azure AI Vision service?
call the Analyze Image function (SDK or equivalent REST method), specifying People as one of the visual features to be returned.
What attributes can be returned in the Facial Attribute analysis?
Head pose (pitch, roll, and yaw orientation in 3D space)
Glasses (NoGlasses, ReadingGlasses, Sunglasses, or Swimming Goggles)
Blur (low, medium, or high)
Exposure (underExposure, goodExposure, or overExposure)
Noise (visual noise in the image)
Occlusion (objects obscuring the face)
Accessories (glasses, headwear, mask)
QualityForRecognition (low, medium, or high)
How can the Face service be provisioned?
You can provision Face as a single-service resource, or you can use the Face API in a multi-service Azure AI Services resource.
Describe the Face detection process in the Face API
When a face is detected by the Face service, a unique ID is assigned to it and retained in the service resource for 24 hours. The ID is a GUID, with no indication of the individual’s identity other than their facial features.
While the detected face ID is cached, subsequent images can be used to compare the new faces to the cached identity and determine if they are similar (in other words, they share similar facial features) or to verify that the same person appears in two images.