Lakehouse Flashcards

(49 cards)

1
Q

What is a lakehouse?

A

A lakehouse presents as a database and is built on top of a data lake using Delta format tables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What capabilities do lakehouses combine?

A

The SQL-based analytical capabilities of a relational data warehouse and the flexibility and scalability of a data lake.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What types of data formats can lakehouses store?

A

All data formats.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the advantage of lakehouses being cloud-based?

A

They can scale automatically and provide high availability and disaster recovery.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What processing engines do lakehouses use?

A

Spark and SQL engines.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the schema-on-read format?

A

Data is organized in a schema-on-read format, meaning the schema is defined as needed rather than having a predefined schema.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What does ACID stand for in the context of lakehouses?

A

Atomicity, Consistency, Isolation, Durability.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are the roles of different users in a lakehouse?

A

Data engineers, data scientists, and data analysts access and use data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the ETL process?

A

Extract, Transform, Load.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What types of data sources can be ingested into a lakehouse?

A

Local files, databases, or APIs.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are Fabric shortcuts?

A

Links to data in external sources, such as Azure Data Lake Store Gen2 or OneLake.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What tools can be used to transform ingested data?

A

Apache Spark with notebooks or Dataflows Gen2.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the purpose of Data Factory pipelines?

A

To orchestrate different ETL activities and land prepared data into the lakehouse.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What familiar tool do Dataflows Gen2 utilize?

A

Power Query.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What can you analyze using a lakehouse?

A

Using SQL.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What can be developed in Power BI using a lakehouse?

A

Reports.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

How is lakehouse access managed?

A

Through workspace roles or item-level sharing.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What are sensitivity labels used for in lakehouses?

A

Data governance features.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

True or False: Item-level sharing is best for granting access for read-only needs.

A

True.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Fill in the blank: Lakehouses support _______ transactions through Delta Lake formatted tables.

21
Q

What is a key benefit of using a lakehouse for analytics?

A

Scalable analytics solution that maintains data consistency.

22
Q

What three items are automatically created in your workspace when you create a new lakehouse?

A

Shortcuts, folders, files, and tables.

The lakehouse serves as a central hub for data management.

23
Q

What does the Semantic model (default) provide for Power BI report developers?

A

An easy data source.

The Semantic model simplifies data representation for reporting.

24
Q

What is the purpose of the SQL analytics endpoint in a lakehouse?

A

Allows read-only access to query data with SQL.

This endpoint enables SQL-based interaction with the lakehouse data.

25
In what two modes can you work with data in the lakehouse?
Lakehouse mode and SQL analytics endpoint mode. ## Footnote Each mode offers different capabilities for managing and querying data.
26
What is the first step in the ETL process for a lakehouse?
Ingesting data into your lakehouse. ## Footnote This step is crucial for preparing data for analysis.
27
List the methods to ingest data into a lakehouse.
* Upload local files * Dataflows Gen2 * Notebooks * Data Factory pipelines ## Footnote Each method has its own use case and benefits.
28
What should you consider when ingesting data to determine your loading pattern?
Whether to load all raw data as files or use staging tables. ## Footnote This decision impacts performance and data processing efficiency.
29
What can Spark job definitions be used for in a lakehouse?
To submit batch/streaming jobs to Spark clusters. ## Footnote This allows for processing large volumes of data efficiently.
30
What is the purpose of shortcuts in a lakehouse?
To integrate data while keeping it stored in external storage. ## Footnote Shortcuts enhance data accessibility across different storage solutions.
31
How are source data permissions and credentials managed when using shortcuts?
They are managed by OneLake. ## Footnote This central management simplifies access control across data sources.
32
What is required for a user to access data through a shortcut to another OneLake location?
The user must have permissions in the target location to read the data. ## Footnote This ensures secure and authorized access to the data.
33
Where can shortcuts be created?
In both lakehouses and KQL databases. ## Footnote This versatility allows for broader data integration options.
34
True or False: Shortcuts appear as a folder in the lake.
True. ## Footnote This structure allows for organized data management within the lakehouse.
35
What is the main role of data transformations in the data loading process?
Most data requires transformations before loading into tables.
36
What tools can be used to transform and load data?
The same tools used to ingest data can also transform and load data.
37
What is a Delta table?
Transformed data can be loaded as a file or a Delta table.
38
Who favors notebooks for data engineering tasks?
Data engineers familiar with different programming languages including PySpark, SQL, and Scala.
39
What interface do Dataflows Gen2 use?
The PowerQuery interface.
40
What do pipelines provide in the ETL process?
A visual interface to perform and orchestrate ETL processes.
41
How complex can pipelines be?
Pipelines can be as simple or as complex as needed.
42
What is required for data to be used after ingestion?
Data must be transformed and loaded.
43
What do Fabric items provide for organizations?
The flexibility needed for every organization.
44
What tools can data scientists use for exploring and training machine learning models?
Notebooks or Data wrangler.
45
What can report developers create using the semantic model?
Power BI reports.
46
What can analysts use the SQL analytics endpoint for?
To query, filter, aggregate, and explore data in lakehouse tables.
47
What is the benefit of combining Power BI with a data lakehouse?
You can implement an end-to-end analytics solution on a single platform.
48
Fill in the blank: After data is ingested, transformed, and loaded, it's ready for _______.
others to use.
49
True or False: Dataflows Gen2 are excellent for developers familiar with SQL only.
False.