Maintain a data analytics solution Flashcards
(9 cards)
You have a Fabric lakehouse that contains a managed Delta table named Product.
You plan to analyze the data by using a Fabric notebook and PySpark.
You load the data to a DataFrame by running the following code.
df = spark.sql(“SELECT * FROM Product”)
You need to display the top 100 rows from the DataFrame.
Which PySpark command should you run?
display(df.limit(100))
You are planning the configuration of a new Fabric tenant.
You need to recommend a solution to ensure that reports meet the following requirements:
Require authentication for embedded reports.
Allow only read-only (live) connections against Fabric capacity cloud semantic models.
Which two actions should you recommend performing from the Fabric admin portal? Each correct answer presents part of the solution.
From Tenant settings, disable Allow XMLA endpoints and Analyze in Excel with on-premises semantic models.
From Tenant settings, disable Publish to web.
You have a Fabric workspace that contains a lakehouse named Lakehouse1.
A user named User1 plans to use Lakehouse explorer to read Lakehouse1 data.
You need to assign a workspace role to User1. The solution must follow the principle of least privilege.
Which workspace role should you assign to User1?
Contributor
(using Lakehouse explorer)
You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 contains a lakehouse, a data pipeline, a notebook, and several Microsoft Power BI reports.
A user named User1 plans to use SQL to access the lakehouse to analyze data. User1 must have the following access:
User1 must have read-only access to the lakehouse.
User1 must NOT be able to access the rest of the items in Workspace1.
User1 must NOT be able to use Spark to query the underlying files in the lakehouse.
You need to configure access for User1.
What should you do?
Share the lakehouse with User1 directly and select Read all SQL Endpoint data.
You have Azure Databricks tables and a Fabric lakehouse.
You need to create a new Fabric artifact to combine data from both architectures. The solution must use data pipelines for the Azure Databricks data and shortcuts for the existing Fabric lakehouse.
What Fabric artifact should you create?
a lakehouse
Only Fabric lakehouses can shortcut to other lakehouses. Fabric data warehouses can use data Pipelines but cannot use shortcuts.
You have a Fabric tenant that contains a workspace named Workspace1. Workspace1 contains two data warehouses named Warehouse1 and Warehouse2. Warehouse1 contains HR data. Warehouse2 contains sales data.
You are analyzing the sales data in Warehouse2 by using the SQL analytics endpoint.
You need to recommend a solution that utilizes a query to combine the sales data from Warehouse2 with the HR data from Warehouse1. The solution must minimize development effort and data movement.
What should you recommend?
Use cross-database querying between Warehouse1 and Warehouse2.
You have a Fabric workspace named Workspace1 that contains a lakehouse named Lakehouse1.
You have write permissions to an Azure Data Lake Storage Gen2 account named storage1 that contains a folder named Folder1.
You plan to delete a shortcut named Shortcut1 that points to a file named File1 stored in Folder1.
You run the delete operation on the following path.
Lakehouse1\Files\Shortcut1
What will occur after you run the delete operation?
Only Shortcut1 will be deleted.
Shortcuts do not perform cascading deletes. However, if the user has write permissions to the shortcut target and specifies the full path of the target, the target will be deleted from the storage account.
You have a Fabric tenant.
Your company has 1 TB of legacy accounting data stored in an Azure Data Lake Storage Gen2 account. The data is queried only once a year for a few ad-hoc reports that submit very selective queries.
You plan to create a Fabric lakehouse or warehouse to store company sales data. Developers must be able to build reports from the lakehouse or warehouse based on the sales data. The developers must also be able to do ad-hoc analysis of the legacy data at the end of each year.
You need to recommend which Fabric architecture to create and the process for integrating the accounting data into Fabric. The solution must minimize administrative effort and costs.
What should you recommend?
Ingest the sales data into the Fabric lakehouse and set up a shortcut to the legacy accounting data in the storage account.