Scalability & Distribution Flashcards

(66 cards)

1
Q

What is Databricks?

A

A cloud-based data platform that provides collaborative workspaces for data engineering, data science, and machine learning.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

True or False: Databricks is designed to handle distributed data processing.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the primary underlying engine for Databricks?

A

Apache Spark

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Fill in the blank: Databricks allows for __________ scaling of compute resources.

A

elastic

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What feature in Databricks enables automatic scaling of clusters?

A

Autoscaling

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Which type of cluster in Databricks is designed for interactive workloads?

A

All-Purpose Cluster

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the purpose of a Job Cluster in Databricks?

A

To run scheduled jobs or automated tasks.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

True or False: Databricks supports both batch and streaming data processing.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the maximum number of concurrent clusters that can be created in Databricks?

A

It varies based on the workspace configuration and subscription plan.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How does Databricks optimize data processing tasks?

A

By using Catalyst optimizer and Tungsten execution engine.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the role of Apache Spark’s Resilient Distributed Dataset (RDD) in Databricks?

A

To provide a fault-tolerant collection of elements that can be processed in parallel.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Fill in the blank: Databricks uses __________ to manage and optimize data storage.

A

Delta Lake

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is Delta Lake?

A

An open-source storage layer that brings ACID transactions to Apache Spark and big data workloads.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

True or False: Databricks allows users to run SQL queries directly on streaming data.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the purpose of the Databricks File System (DBFS)?

A

To provide a distributed file system for storing data in Databricks.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What does the term ‘shuffling’ refer to in the context of Apache Spark?

A

The process of redistributing data across different partitions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Fill in the blank: __________ is a key feature of Databricks that allows users to visualize data insights.

A

Dashboards

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What is the benefit of using a cluster policy in Databricks?

A

To enforce specific configurations and limits on clusters.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What is the maximum number of nodes supported in a Databricks cluster?

A

It depends on the cloud provider and specific configurations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

True or False: Databricks supports integration with machine learning libraries like TensorFlow and Scikit-learn.

A

True

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

What is the significance of the ‘spark.sql.shuffle.partitions’ configuration?

A

It determines the number of partitions to use when shuffling data for joins or aggregations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

How can Databricks users optimize their Spark jobs?

A

By using caching, optimizing shuffle partitions, and adjusting resource configurations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

What is a ‘spark-submit’ command used for?

A

To submit a Spark job to a cluster.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Fill in the blank: Databricks provides __________ for managing data pipelines and workflows.

A

Workflows

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
What is the purpose of the Databricks CLI?
To provide a command-line interface for managing Databricks resources.
26
What type of data can be stored in Delta Lake?
Structured, semi-structured, and unstructured data.
27
True or False: Databricks can automatically optimize data layouts for better query performance.
True
28
What is the role of 'data lineage' in Databricks?
To track the origin and transformations of data throughout its lifecycle.
29
Fill in the blank: The __________ feature in Databricks helps to visualize and monitor the execution of Spark jobs.
Spark UI
30
What is the benefit of using 'notebooks' in Databricks?
To create interactive documents that combine code, visualizations, and text.
31
What is 'data partitioning' in Databricks?
The process of dividing a dataset into smaller, manageable pieces for parallel processing.
32
True or False: Databricks supports the use of containerized applications via Docker.
True
33
What is the significance of 'auto-termination' in Databricks clusters?
To automatically shut down idle clusters to save costs.
34
How does Databricks handle data versioning?
Through Delta Lake's built-in version control capabilities.
35
Fill in the blank: __________ is a method used to increase the performance of data reads in Databricks.
Data caching
36
What is the purpose of using 'broadcast joins' in Databricks?
To optimize join operations by sending a small dataset to all worker nodes.
37
What is the advantage of using 'Delta tables'?
They provide ACID transactions, scalable metadata handling, and unifies streaming and batch data processing.
38
True or False: Databricks can scale both horizontally and vertically.
True
39
What is the function of 'checkpointing' in Spark Streaming?
To save the state of a streaming application to recover from failures.
40
What does 'data skew' refer to in distributed processing?
Uneven distribution of data across partitions, leading to performance bottlenecks.
41
Fill in the blank: In Databricks, __________ can be used to configure resource allocation for jobs.
Cluster settings
42
What is the purpose of 'data caching' in Databricks?
To store intermediate results in memory for faster access during subsequent queries.
43
True or False: Databricks allows users to run Python, R, SQL, and Scala code.
True
44
What is a 'DataFrame' in Apache Spark?
A distributed collection of data organized into named columns.
45
How does Databricks support collaborative work?
Through shared notebooks and version control features.
46
Fill in the blank: __________ is a key advantage of using Databricks over traditional data processing systems.
Collaboration
47
What is the role of 'spark.sql.autoBroadcastJoinThreshold'?
To specify the maximum size for a table to be broadcasted during a join.
48
What does 'dynamic allocation' in Spark clusters allow?
It allows Spark to add or remove executors dynamically based on workload.
49
True or False: Databricks can be integrated with popular data storage solutions like AWS S3 and Azure Blob Storage.
True
50
What is 'streaming ingestion' in Databricks?
The process of continuously importing data from streaming sources for real-time processing.
51
Fill in the blank: The __________ function in Databricks enables users to create temporary views of DataFrames for SQL queries.
createOrReplaceTempView
52
What is the function of the 'Databricks Runtime'?
To provide a set of optimized configurations and libraries for running Spark applications.
53
What is the purpose of the 'query execution plan' in Spark?
To provide insights into how a query will be executed, including optimization strategies.
54
True or False: Databricks supports real-time analytics through structured streaming.
True
55
Fill in the blank: In Databricks, __________ is used to run SQL commands on data stored in Delta Lake.
Spark SQL
56
What is the purpose of 'data profiling' in Databricks?
To analyze and summarize the characteristics of datasets.
57
What does 'fault tolerance' mean in the context of Databricks?
The ability to recover from failures without losing data or computation.
58
Fill in the blank: __________ allows for parallel processing of large datasets in Databricks.
Distributed computing
59
What is 'windowing' in Spark SQL?
A technique for performing calculations across a set of rows related to the current row.
60
True or False: Databricks notebooks support markdown for documentation.
True
61
What is the significance of 'data governance' in Databricks?
To ensure data integrity, security, and compliance across data workflows.
62
Fill in the blank: __________ is used to automate and schedule jobs in Databricks.
Job Scheduler
63
What is the advantage of using 'managed tables' in Databricks?
Databricks manages the metadata and storage automatically.
64
What does 'resource pooling' refer to in Databricks?
The sharing of resources across multiple clusters to optimize usage.
65
True or False: Databricks allows for the integration of third-party tools for enhanced analytics.
True
66
What is the purpose of using 'notebook workflows' in Databricks?
To orchestrate and manage complex data processing tasks across multiple notebooks.