DP-203 Azure Data Engineer

Lakukan tugas rumah & ujian kamu dengan baik sekarang menggunakan Quizwiz!

What variables must you initialize in order to write a PowerShell script to copy an archive blob to a new blob within the same storage account?

1. $rgName - your resource group name 2. $accountName - your storage account name 3. $srcContainerName and $destContainerName - your container name 4. $srcBlobName and $destBlobName - your blob names

What are the advantages of using columnstore indexes?

1. 10x performance in query performance 2. 10x data compression

Which functions in Azure Stream Analytics can be used for anomaly detection?

1. AnomalyDetection_SpikeAndDip 2. AnomalyDetection_ChangePoint

What are the ACID guarantees for a transaction?

1. Atomicity - means a transaction must execute exactly once and must be atomic; either all of the work is done, or none of it is. Operations within a transaction usually share a common intent and are interdependent 2. Consistency - ensures that the data is consistent both before and after the transaction 3. Isolation - ensures that one transaction is not impacted by another transaction 4. Durability - means that the change made due to the transaction are permanently saved in the system. Committed data is saved by the systems so that even in the event of a failure and system restart, the data is available in its correct state

What storage layers are supported by Stream Analytics?

1. Azure Blob Storage 2. Azure SQL Database

What are the four data services that make up Azure Storage?

1. Azure Blobs 2. Azure Files 3. Azure Queues 4. Azure Tables

Which two Azure services do NOT provide row level security?

1. Azure Data Explorer 2. Cosmos DB

Which two Azure batch processing solutions do not support auto-scaling?

1. Azure Data Lake Analytics 2. Azure Synapse

Which batch processing solutions do not support in-memory caching of data?

1. Azure Data Lake Analytics 2. HDInsight with Hive

Which batch processing solutions do not support Row-level security?

1. Azure Data Lake Analytics 2. HDInsight with Spark 3. Azure Databricks

Which batch processing solutions do not support dynamic data masking?

1. Azure Data Lake Analytics 2. HDInsight with Spark 3. Azure Databricks

Which Azure services provide the facility of dynamic data masking?

1. Azure Synapse 2. HDInsight

Which batch processing solutions do not support query from external relational tables?

1. Azure Synapse 2. HDInsight with Hive 3. HDInsight with LLAP

What does Azure Site Recovery manage replication for?

1. Azure VMs replicating between Azure regions 2. On-premises VMs, Azure Stack VMs, and physical servers

What three logging levels are available in ADF?

1. Basic 2. None 3. Verbose

Using an external table for export scenario is limited to which formats?

1. CSV 2. TSV 3. JSON 4. Parquet

What actions can the ALTER FUNCTION not perform?

1. Change a scalar-valued function to a table-valued function, or vice-versa 2. Change an inline function to a multi-statement function or vice-versa 3. Change a Transact-SQL function to a CLR function, or vice-versa

What are the two options for rehydrating a blob that is stored in the archive tier?

1. Copy an archived blob to an online tier: You can rehydrate an archived blob by copying it to a new blob in the hot or cool tier with the Copy Blob or Copy Blo from URL operation. (Microsoft recommends this approach in most scenarios) 2. Change a blob's access tier to an online tier: You can rehydrate an archived blob to hot or cool by changing its tier using the Set Blob Tier operation

What are the steps to creating an Azure AD application and service principal to mount or directly access data in an ADLS Gen2 storage account?

1. Create an Azure AD application (this will create an associated service principal used to access the storage account) 2. Create an Azure Key Vault-backed secret scope. (the secret scope will securely store the client secret associated with the Azure AD application) 3. Save the client secret associated with the Azure AD application in the Azure key vault 4. Assign roles to the application to provide the service principal with the required permissions to access the ADLS Gen2 storage account 5. Create one or more containers inside the storage account. 6. Authenticate and access the ADLS Gen2 storage account through a mount point or direct access

What are the considerations for an incremental copy pattern for a data lake migration?

1. Cutover from Gen1 to Gen2 for all workloads at the same time 2. Expect downtime during cutover period only 3. Ideal for pipelines where all apps upgraded at one time, but the data copy requires more time

What are the considerations for a life and shift pattern for a data lake migration?

1. Cutover from Gen1 to Gen2 for all workloads at the same time 2. Expect downtime during the migration and the cutover period 3. Ideal for pipelines that can afford downtime and all apps can be upgraded at one time

What five connection properties need to be collected and stored as environment variables for deployment in IDE?

1. DATABRICKS_ADDRESS 2. DATABRICKS_API_TOKEN 3. DATABRICKS_CLUSTER_ID 4. DATABRICKS_PORT 5. DATABRICKS_ORG_ID

What services does Azure Databricks provide diagnostic logs for?

1. DBFS 2. Clusters 3. Pools 4. Accounts 5. Jobs 6. Notebook 7. SSH 8. Workspace 9. Secrets 10. Databricks SQL 11. SQL Permissions 12. Repos

In which scenarios should you use Azure Stream Analytics?

1. Dashboards for data visualization 2. Real-time alerts from temporal and spatial patterns or anomalies 3. Extract, Transform, Load (ETL) 4. Event Sourcing pattern 5. IoT edge

What three factor are used to determine the number of storage accounts to create?

1. Data diversity 2. Cost sensitivity 3. Tolerance for management overhead

What are the logical components of a batch processing architecture?

1. Data storage (Azure Storage Blob Containers or Azure Data Lake Store) 2. Batch processing (U-SQL, Hive, Pig, or Spark) 3. Analytical Data Store (Azure Synapse Analytics, Spark SQL, HBase, or Hive) 4. Analysis and reporting (Azure Analysis Services, Power BI, or Excel) 5. Orchestration (Azure Data Factory, Oozie, or Sqoop)

What facilities in a data flow tool allow you to protect against schema drift?

1. Define sources that have mutable field names, data types, values, and sizes 2. Define transformation parameters that can work with data patterns instead of hard-coded fields and values 3. Define expressions that understand patterns to match incoming fields, instead of using named fields

What should you do to ensure your table service solution is read-efficient?

1. Design for querying in read-heavy applications 2. Specify both PartitionKey and RowKey in your queries 3. Consider storing duplicate copies of entities 4. Consider denormalizing your data 5. Use compound key values 6. Use query projection

What should you do to ensure your table service is write-efficient?

1. Do not create hot partitions 2. Avoid spikes in traffic 3. Don't necessarily create a separate table for each type of entity 4. Consider the maximum throughput you must achieve

What are the guiding principles recommended to update the statistics during the load process?

1. Ensure that each loaded table has at least one statistic object updated. This process updates the table size (row count and page count) information as part of the statistics update 2. Focus on columns participating in JOIN, GROUP BY, ORDER BY, and DISTINCT clauses 3. Consider updating "ascending key" columns such as transaction dates more frequently because these values won't be included in the statistics histogram 4. Consider updating static distribution columns less frequently 5. Remember, each statistic object is updated in sequence. Simply implementing UPDATE STATISTICS <TABLE_NAME> isn't always ideal, especially for wide tables with lots of statistics objects

What are the requirements for switching partitions between two tables?

1. Ensure that the partitions alight on their respective boundaries 2. Ensure that the table definitions match

Which Azure resources can be used for storing diagnostic logs for Databricks?

1. Event hub 2. Storage account 3. Azure Log Analytics

Which GROUP BY options are supported in dedicated SQL pool?

1. GROUP BY with ROLLUP 2. GROUPING SETS 3. GROUP BY with CUBE

What are the considerations for a dual pipeline pattern for a data lake migration?

1. Gen1 and Gen2 pipelines run side-by-side 2. Supports zero downtime 3. Ideal in situations where your workloads and applications can't afford any downtime, and you can ingest into both storage accounts

What are the steps to run a CI/CD pipeline in ADF?

1. Go to the Pipelines page. Then choose the action to create a new pipeline 2. Select Azure Repos Git as the location of your source code 3. When the list of repositories appears, select your repository 4. As you setup your pipeline, select Existing Azure Pipelines YAML file 5. Run the pipeline

In order to balance parallel processing, what properties should a distribution column have?

1. Has many unique values 2. Does not have NULLs or has only a few NULLs 3. Is not a date column (all data from the same date lands in the same distribution)

What are the considerations for using the bi-directional sync pattern for a data lake migration?

1. Ideal for complex scenarios that involve a large number of pipelines and dependencies where a phased approach might make more sense 2. Migration effort is high, but it provides side-by-side support for Gen1 and Gen2

What are the solutions to resolving data skew problems?

1. Improve table partitioning a. Option 1: Filter the skewed key value in advance b. Option 2: Pick a different partition or distribution key c. Option 3: Add more partition or distribution keys d. Option 4: Use round-robin distribution 2. Improve the query plan a. Option 1: Use the CREATE STATISTICS statement b. Option 2: Use SKEWFACTOR c. Option 3: Use ROWCOUNT 3. Improve the user-defined reducer and combiner a. Option 1: Use a recursive reducer, if possible b. Option 2: Use row-level combiner mode, if possible

What four function types are supported by Azure Stream Analytics?

1. JavaScript user-defined functions 2. JavaScript user-defined aggregates 3. C# user defined functions (with Visual Studio) 4. Azure Machine Learning

What parameters can be specified for a job?

1. Job Constraints: specified to set the maximum amount of time a job can be active or running (maxWallClockTime) 2. Job Preparation Task: a job preparation task is run the first time a task in run for a job on a node 3. Job Release Task: can only be specified if a job preparation task is configured, when a job is being terminated, the job release task is run on the each of pool nodes where a job preparation task was run

Azure role-based control (RBAC) is the authorization technique that can be used for managing access to Azure resources, which four top-level classifications can you assign roles to grant access

1. Managed identities 2. Users 3. Groups 4. Service Principals

What are the three layers of an HDInsight cluster?

1. Physical Layer 2. YARN Layer 3. Workload Layer

What are some limitations in Azure synapse SQL that may block you from enabling auditing?

1. Premium storage is currently not supported 2. Hierarchical namespace for Azure Data Lake Storage Gen2 storage account is currently not supported 3. Enabling auditing on a paused Azure Synapse is not supported. To enable auditing, resume Azure Synapse 4. Auditing for Azure Synapse SQL pools supports default audit action groups only

What are the best practices when using Delta Lake?

1. Provide data location hints: if you expect a column to be commonly used in query predicates and if that column has high cardinality, then use Z-ORDER BY. Delta Lake automatically lays out the data in the files based on the column values and use the layout information to skip irrelevant data while querying 2. Compact files: a large number of small files should be rewritten into a smaller number of larger files on a regular basis 3. Replace the content or schema of a table: if you don't need to change the table schema, you can delete data from a Delta table and insert your new data, or update the table to fix incorrect values

What can SQL database auditing be used for?

1. Retain an audit trail of select events. You can define categories of database actions to be audited. 2. Report on database activity. You can pre-configure reports and a dashboard to get started quickly with activity and event reporting 3. Analyze reports. You can find suspicious events, unusual activity, and trends

What can an Azure integration runtime do?

1. Run Data Flows in Azure 2. Run copy activity between cloud data stores 3. Dispatch select transform activities in public network

What can an Azure self-hosted integration runtime do?

1. Run copy activity between a cloud data stores and a data store in private network 2. Dispatch select transform activities against compute resources in on-premises or Azure Virtual Network

In Azure Synapse, turning on resource logs collects what data?

1. SQLSecurityAuditEvents 2. DmsWorkers 3. ExecRequests 4. SQLRequests 5. Waits 6. All Metrics

What are the steps to incrementally load from multiple tables in SQL Server to a database in Azure SQL database using the Azure portal?

1. Select the watermark column 2. Prepare a data store to store the watermark value 3. Create a pipeline with the following activities a. Create a ForEach loop that iterates through a list of source tables b. Create two lookup activities c. Create a copy activity that copies rows from the source data store d. Create a stored procedure to update the watermark value for the next pipeline run

What are the steps to setup access control for your Azure Synapse workspace?

1. Set up security groups 2. Prepare your ADLS Gen2 storage account 3. Create and configure your Azure Synapse Workspace 4. Grant the workspace MSI access to the default storage container 5. Grant Synapse administrators the Azure Contributor role on the workspace 6. Assign SQL Active Directory Admin role 7. Grant access to SQL pools 8. Add users to security groups 9. Network Security 10. Completion

What are the two Rehydration priority options

1. Standard priority: the rehydration request will be processed in the order it was received and may take up to 15 hours 2. High priority: The rehydration request will be prioritized over standard priority requests and may complete in under one hour for objects under 10 GB in size

Which targets can diagnostic logs be routed to from Azure Monitor?

1. Storage Account 2. Event Hub 3. Log Analytics (use this when you want to write complex queries on a rich set of metrics and/or you want to monitor across multiple factories)

What are the settings that are defined by a storage account?

1. Subscription 2. Location 3. Performance 4. Replication 5. Access tier 6. Secure transfer protocol 7. Virtual networks

What are the three ways to create a system-versioned temporal table with regards to how the history table is specified?

1. Temporal table with an anonymous history table (you specify the schema of the current table and let the system create a corresponding history table with auto-generate name) 2. Temporal table with a default history table (you specify the history table schema name and table name and let the system create a history table in that schema) 3. Temporal table with a user-defined history table created beforehand (you create a history table that fits best your needs and then reference that table during temporal table creation)

In which scenarios might you need to provide a public endpoint connection with Azure SQL Managed instance?

1. The managed instance must integrate with multi-tenant-only platform-as-a-service (PaaS) offerings 2. You need higher throughput of data exchange than is possible when you're using a VPN 3. Company policies prohibit PaaS inside corporate networks

How do you avoid data movement during a join?

1. The tables involved in the join must be hashed distributed on one of the columns participating in the join 2. The data types of the join columns must match between both tables 3. The columns must be joined with an equals operator 4. The join type may not be a CROSS JOIN

What are the methods to accessing drifted columns in your data flow?

1. Use the byPosition and byName expressions to explicitly reference a column by name or position number 2. Add a column pattern in a Derived Column or Aggregate transformation to match on any combination of name, stream, position, origin, or type 3. Add rule-based mapping in a Select or Sink transformation to match drifted columns to columns aliases via a pattern

What are the three common serialization languages that you will encounter?

1. XML (extensible markup language) - flexible and can express complex data easily, however it tends to be more verbose making it larger to store, process, or pass over a network 2. JSON (Javascript Object Notation) - has lightweight specification and relies on curly braces to indicate data structure. It is less verbose than XML 3. YAML - data structure is defined by line separation and indentation, and reduces the dependency on structural characters like parentheses, commas and brackets

In which scenarios should you NOT use Azure Stream Analytics?

1. You want to write UDF, UDAs, and custom deserializers in a language other than JavaScript or C# 2. Your solution is in a multi-cloud or on-premises environment

In Spark 2.1, what are the supported compression types of Parquet data types?

1. none 2. snappy 3. gzip 4. lzo

In Spark 2.4, what are the supported compression types?

1. none 2. uncompressed 3. snappy 4. gzip 5. lzo 6. brotli 7. lz4 8. zstd

How long is Pipeline-run data stored in Data Factory?

45 days

What is Hive?

A SQL like language that is supported in most Hadoop distributions, including HDInsight. It can be used to process data from any HDFS-compatible store, including Azure blob storage and Azure Data Lake Store. In addition, Hive offers a database architecture that is conceptually similar to that of a typical relational database management system.

What is schema drift?

A case where your sources often change metadata. Fields, columns, and types can be added, removed, or changed on the fly

In Databricks, what does the job latency visualization show?

A coarse view on the overall performance of a job. Displays the job execution duration from start to completion.

What is a storage account?

A container that groups a set of Azure Storage services together. Only data services from Azure Storage can be included in a storage account

What is a job ?

A group of one or more tasks, with the tasks actually specifying the command lines to be run.

What is HBase?

A low-latency NoSQL store that offers high-performance, flexible option for querying structured and semi-structured data

What is Azure Synapse Analytics?

A managed service based on SQL Server database technologies and optimized to support large-scale data warehousing workloads

What is U-SQL?

A query processing language used by Azure Data Lake Analytics. It combines the declarative nature of SQL with the procedural extensibility of C#, and takes advantage of parallelism to enable efficient processing of data at a massive scale

What is a transaction?

A transaction is a logical group of database operations that execute together

Which logs provide insight into operations on each Azure resource in the subscription from the outside (the management plane) in addition to updates on Service Health events

Activity logs

What is Spark SQL?

An API built on Spark that supports the creation of dataframes and tables that can be queried using SQL Syntax

Which block blob storage tier has a minimum storage duration of 180 days and a latency in hours?

Archive tier

What does Combiner mode do?

Attempts to distribute very big skewed-key value sets to different vertices

Which logs contain the history of sign-in activity and audit trail of changes made in the Azure Active Directory for a particular tenant?

Azure Active Directory logs

Which Azure analytical data store does not have SQL language support?

Azure Analysis Services

Which Azure service is a fast, fully managed data analytics service for real-time analysis of large volumes of data?

Azure Data Explorer

Which batch processing solution does not support firewalls?

Azure Databricks

Where must customer-managed keys be stored?

Azure Key Vault or Azure Key Vault Managed Hardware Security Model (HSM)

Which Azure service provides base-level infrastructure metrics and logs for most Azure Services?

Azure Monitor

What is the best way to perform ETL and move data from Azure Synapse to Azure Databricks?

Azure Synapse Connector

Currently, which type of data base is not supported as a sink for the copy activity?

Azure database for Postgres

What T-SQL statement allows you to create an external table in a dedicated SQL pool or serverless SQL pool?

CETAS

Which T-SQL statements allows you to specify both the distribution of the table data as well as the table structure type?

CTAS

What should you do if you receive the following error message in a Spark job: "Serialized task is too large"

Call parallelize with a large list or convert a large R DataFrame to a Spark DataFrame

Which block blob storage tier has a minimum storage duration of 30 days and a latency in milliseconds?

Cool tier

What is unstructured data?

Data that is ambiguous and often delivered in files such as photos or videos.

What is semi-structured data?

Data that is not stored in a relational format, as the fields do not neatly fit into tables, rows, and columns. The data contains tags that make the organization and hierarchy of the data apparent. Semi-structured data is also referred to as non-relational or NoSQL data. The expression and structure of the data in this style is defined by a serialization language

What is the distributed file systems mounted into an Azure Databricks workspace and available on Azure Databricks cluster?

Databricks File System (DBFS)

What service is used to connect an IDE, notebook server, or other custom application to Azure Databricks clusters?

Databricks connect

What type of read causes a transaction reads data that has not yet been committed

Dirty Read

Which Azure tools should be used for ingesting streaming data?

HDInsight Storm or Azure Stream Analytics

Which batch processing solution does not support Auditing?

HDInsight with Spark

Which batch processing solution does not support Azure Active Directory?

HDInsight with Spark

In which sharding strategy is data distributed across the shards in a way that achieves a balance between size of each shard and the average load that each shard will encounter

Hash strategy

Which window functions model scheduled overlapping windows, jumping forward in time by a fixed period

Hopping Window

Which block blob storage tier has no minimum storage duration and a latency in milliseconds?

Hot Tier

What kind of data redundancy does Azure Data Factory have?

In all regions (except Brazil South and Southeast Asia), Azure Data Factory data is stored and replicated in the paired region to protect against metadata loss. During regional data center failures, Microsoft may initiate a regional failover of your Data Factory instance. In most cases, no action is required on the users part.

Which real time message ingestion technology should you use if you need two-way communication or you need to manage access for individual access for individual devices and be able to revoke access to a specific device?

IoT Hub

In which architecture is the serving layer subdivided into a speed serving layer, which stores data that has been processed incrementally, and a batch serving layer, which contains the batch-processed output.

Lambda architecture

In which sharding strategy does the sharding logic implement a map that routes a request for data to the shard that contains the data using the shard key

Lookup strategy

Which sharding strategy offers more control over the way that shards are configured and used?

Lookup strategy

Are block predicates supported in Azure Synapse?

No

Can you disable Synapse Link at the account level after it is enabled?

No, you must either delete and re-create the Azure CosmosDB account or open a support ticket with Microsoft

What type of read causes a transaction reads the same row twice and gets a different value each time?

Non Repeatable Read

What type of read causes two same queries are executed but the rows retrieved are different?

Phantom Read

Which block blob storage tier has no minimum storage duration and a latency in single-digit milliseconds?

Premium Performance Tier

In which sharding strategy do groups of related items get grouped in the same shard, and orders them by shard key

Range strategy

Which sharding strategy is easy to implement and works well with range queries because they can often fetch multiple data items from a single shard in a single operation?

Range strategy

Which SQL isolation level guarantees that any data read is committed at the moment it is read. Thus not allowing dirty reads. The transaction holds a read or write lock on the current row, and thus prevent other transactions from reading, updating, or deleting it?

Read Committed

Which SQL isolation level allows one transaction to read not yet committed changes made by another transaction, thereby allowing dirty reads?

Read Uncommitted

Which SQL isolation level is the most restrictive and holds read locks on all rows it references and writes locks on all rows it inserts, updates, or deletes.

Repeatable Read

Which logs provide insight into operations that were performed within an Azure resource (the data plane)

Resource logs

In RBAC, what is a collection of permissions that lists actions that can be performed, such as read, write, and delete

Role definition

In RBAC, what is a set of resources that the access applies to

Scope

In RBAC, what is an object that represents a user, group, service principal, or managed identity that is requesting access to Azure resources?

Security principal

Which SQL isolation level is the highest isolation level and execution is guaranteed to be serializable. Serializable execution is defined to be an execution of operations in which concurrently executing transactions appears to be serially executing

Serializable

Which window function clusters together events that arrive at similar times, filtering out periods of time where there is no data?

Session Window

Which window function generates events for points in time when the content of the window actually changed?

Sliding Window

Which window function groups events by identical timestamp values

Snapshot Window

What is structured data?

Sometimes referred to as relational data, is data that adheres to a strict schema, so all of the data has the same fields or properties

What Apache Spark API lets you express computation on streaming data in the same way you express a batch computation on static data?

Structured Streaming

When are statistics automatically created in Synapse SQL?

The dedicated SQL pool engine will analyze incoming user queries for missing statistics when the database AUTO_CREATE_STATISTICS option is set to ON. If statistics are missing, the query optimizer creates statistics on individual columns in the query predicate or join condition

In Databricks, what does the stage latency visualization show?

The latency of each stage per cluster, per application, and per individual stage. (Useful for identifying a particular stage that is running slowly)

In Databricks, what does the value assigned to spark.driver.maxResultSize define?

The maximum size (in bytes) of the serialized results for each Spark action. This value cannot be negative

What is a Deployment Model for a storage account?

The system Azure uses to organize your resources

Which window function segments a data stream into a contiguous series of fixed-size, non-overlapping time segments and operates against them?

Tumbling Window

Which SCD type always reflects the latest values and when cahnges in source data are detected, the dimension table data is overwritten?

Type 1 SCD

Which SCD type supports versioning of dimensions? In this case, the dimension table must use a surrogate key to provide a unique reference to a version of the dimension number. It also includes columns that define the date range validity of the version (ex: StartDate & EndDate) and possibly a current record flag

Type 2 SCD

Which SCD type supports storing two versions of a dimension member as separate columns

Type 3 SCD

How do you enable schema drift in the sink?

Under sink, check the box "Allow schema drift"

How do you enable schema drift in source?

Under source settings, check the box "Allow Schema Drift"

How to access data in an Azure Data Lake Gen2 (ADLS Gen2) account with Azure Databricks?

Use OAuth 2.0 with an Azure Active Directory (Azure AD) application service principal for authentication

What type of Databricks cluster can be created using the UI, CLI, or REST API and can be manually terminated and restart?

all-purpose cluster

What is LAG?

an analytic operator that allows us to look up a previous event part of an event stream. It is majorly used in finding the growth or change or a particular variable, which is the temperature of the Panel

With RLS, what do block predicates do?

block predicates explicitly block write operations that violate the predicate

What attributes does the Copy activity support preserving?

contentType, contentLanguage, contentEncoding, contentDisposition, cacheControl

Zone-redundant storage (ZRS)

copies your data synchronously across three Azure availability zones in the primary region. For applications requiring high availability, Microsoft recommends using ZRS in the primary region and also replicating to a secondary region

Locally redundant storage (LRS)

copies your data synchronously three times within a single physical location in the primary region. This is the least expensive replication option, but it is not recommended for applications requiring high availability or durability

What type of Databricks cluster provides fine-grained sharing for maximum resource utilization and minimum query latencies?

high concurrency clusters

What is a shell script that runs during startup of each cluster node before the Apache Spark driver or worker JVM starts?

init script

What type of Databricks cluster is created by job scheduler when you run a job and is automatically terminated when the job is complete?

job cluster

In RBAC, what is the process of attaching a role definition to a user, group, service principle, or managed identity at a particular scope for the purpose of granting access?

role assignment

What type of encryption does Azure Storage use for all storage accounts?

server-side encryption (SSE) using 256-bit AES encryption

With RLS, what do filer predicates do?

silently filter the rows available to read operations (SELECT, UPDATE, and DELETE)

What type of Databricks cluster consists of an Apache Spark driver and no Spark workers?

single node cluster

What type of Databricks cluster is recommended for a single user that can run workloads developed in Python, SQL, R, and Scala?

standard cluster

What is LIMIT DURATION?

used to restrict the interval or time considered while computing a query


Set pelajaran terkait

CH17, CH18, Ch 19, Ch 20, Ch 21, Ch 22, Ch 23, Ch 24, Ch 25, Ch 26, Ch 27

View Set

NUR3010: CHAPTER 2- PrepU Quizzes

View Set