DP 200 (No Drag and Drop)

Réussis tes devoirs et examens dès maintenant avec Quizwiz!

You have an Azure SQL database named DB1 that contains a table named Table1. Table1 has a field named Customer_ID that is varchar(22).You need to implement masking for the Customer_ID field to meet the following requirements:✑ The first two prefix characters must be exposed.✑ The last four prefix characters must be exposed.✑ All other characters must be masked.Solution: You implement data masking and use an email function mask.Does this meet the goal?

Must use Custom Text data masking, which exposes the first and last characters and adds a custom padding string in the middle.

You need to load the data from the Azure Data Lake Gen 2 storage account into the Azure SQL Data Warehouse.Solution:1. Create an external data source pointing to the Azure storage account2. Create a workload group using the Azure storage account name as the pool name3. Load the data using the INSERT"¦SELECT statementDoes the solution meet the goal?

NO

You develop a data ingestion process that will import data to a Microsoft Azure SQL Data Warehouse. The data to be ingested resides in parquet files stored in anAzure Data Lake Gen 2 storage account.You need to load the data from the Azure Data Lake Gen 2 storage account into the Azure SQL Data Warehouse.Solution:1. Use Azure Data Factory to convert the parquet files to CSV files2. Create an external data source pointing to the Azure storage account3. Create an external file format and external table using the external data source4. Load the data using the INSERT"¦SELECT statementDoes the solution meet the goal?

NO External File Format supported by Polybase are 1. Delimited text 2. Hive RCFile 3. Hive ORC 4. Parquet

You have an Azure SQL database named DB1 that contains a table named Table1. Table1 has a field named Customer_ID that is varchar(22).You need to implement masking for the Customer_ID field to meet the following requirements:✑ The first two prefix characters must be exposed.✑ The last four prefix characters must be exposed.✑ All other characters must be masked.Solution: You implement data masking and use a random number function mask.

NO Must use Custom Text data masking, which exposes the first and last characters and adds a custom padding string in the middle.

You have an Azure SQL database named DB1 that contains a table named Table1. Table1 has a field named Customer_ID that is varchar(22).You need to implement masking for the Customer_ID field to meet the following requirements:✑ The first two prefix characters must be exposed.✑ The last four prefix characters must be exposed.✑ All other characters must be masked.Solution: You implement data masking and use a credit card function mask.Does this meet the goal?

NO Must use Custom Text data masking, which exposes the first and last characters and adds a custom padding string in the middle.References:https://docs.microsoft.com/en-us/azure/sql-database/sql-database-dynamic-data-masking-get-started

You need to load the data from the Azure Data Lake Gen 2 storage account into the Azure SQL Data Warehouse.Solution:1. Create an external data source pointing to the Azure storage account2. Create an external file format and external table using the external data source3. Load the data using the INSERT"¦SELECT statementDoes the solution meet the goal?

NO You load the data using the CREATE TABLE AS SELECT statement.

Solution: You create a High Concurrency cluster for each data scientist, a High Concurrency cluster for the data engineers, and a Standard cluster for the jobs.Does this meet the goal?

No

Solution: You create a Standard cluster for each data scientist, a High Concurrency cluster for the data engineers, and a High Concurrency cluster for the jobs.Does this meet the goal? A. Yes B. No

No

You need to load the data from the Azure Data Lake Gen 2 storage account into the Azure SQL Data Warehouse.Solution:1. Create a remote service binding pointing to the Azure Data Lake Gen 2 storage account2. Create an external file format and external table using the external data source3. Load the data using the CREATE TABLE AS SELECT statementDoes the solution meet the goal?

No

You need to load the data from the Azure Data Lake Gen 2 storage account into the Azure SQL Data Warehouse.Solution:1. Create an external data source pointing to the Azure storage account2. Create a workload group using the Azure storage account name as the pool name3. Load the data using the CREATE TABLE AS SELECT statementDoes the solution meet the goal?

No

You have an Azure Storage account that contains 100 GB of files. The files contain text and numerical values. 75% of the rows contain description data that has an average length of 1.1 MB.You plan to copy the data from the storage account to an Azure SQL data warehouse.You need to prepare the files to ensure that the data copies quickly.Solution: You modify the files to ensure that each row is more than 1 MB.Does this meet the goal? A. Yes B. No

Not Sure

You have an Azure Storage account that contains 100 GB of files. The files contain text and numerical values. 75% of the rows contain description data that has an average length of 1.1 MB.You plan to copy the data from the storage account to an Azure SQL data warehouse.You need to prepare the files to ensure that the data copies quickly.Solution: You modify the files to ensure that each row is less than 1 MB.Does this meet the goal? A. Yes B. No

Not sure

Contoso, Ltd. plans to configure existing applications to use Azure SQL Database.When security-related operations occur, the security team must be informed.You need to configure Azure Monitor while minimizing administrative effort.Which three actions should you perform? Each correct answer presents part of the solution.NOTE: Each correct selection is worth one point. A. Create a new action group to email [email protected]. B. Use [email protected] as an alert email address. C. Use all security operations as a condition. D. Use all Azure SQL Database servers as a resource. E. Query audit log entries as a condition.

Not sure Correct Answer: ACDReferences:https://docs.microsoft.com/en-us/azure/azure-monitor/platform/alerts-action-rules

Solution: You create a Standard cluster for each data scientist, a High Concurrency cluster for the data engineers, and a Standard cluster for the jobs.

Yes

You need to load the data from the Azure Data Lake Gen 2 storage account into the Azure SQL Data Warehouse.Solution:1. Create an external data source pointing to the Azure Data Lake Gen 2 storage account2. Create an external file format and external table using the external data source3. Load the data using the CREATE TABLE AS SELECT statementDoes the solution meet the goal?

Yes

A project requires the deployment of resources to Microsoft Azure for batch data processing on Azure HDInsight. Batch processing will run daily and must:✑ Scale to minimize costs✑ Be monitored for cluster performanceYou need to recommend a tool that will monitor clusters and provide information to suggest how to scale.Solution: Monitor clusters by using Azure Log Analytics and HDInsight cluster management solutions.Does the solution meet the goal? A. Yes B. No

A

You are creating a new notebook in Azure Databricks that will support R as the primary language but will also support Scola and SQL.Which switch should you use to switch between languages? A. %<language> B. \\[<language>] C. \\(<language>) D. @<Language>

A. %<language>

You have an Azure Storage account and an Azure SQL data warehouse by using Azure Data Factory. The solution must meet the following requirements:✑ Ensure that the data remains in the UK South region at all times.✑ Minimize administrative effort.Which type of integration runtime should you use? A. Azure integration runtime B. Self-hosted integration runtime C. Azure-SSIS integration runtime

A. Azure integration runtime

You need to develop a pipeline for processing data. The pipeline must meet the following requirements:✑ Scale up and down resources for cost reduction✑ Use an in-memory data processing engine to speed up ETL and machine learning operations.✑ Use streaming capabilities✑ Provide the ability to code in SQL, Python, Scala, and R✑ Integrate workspace collaboration with GitWhat should you use? A. HDInsight Spark Cluster B. Azure Stream Analytics C. HDInsight Hadoop Cluster D. Azure SQL Data Warehouse E. HDInsight Kafka Cluster F. HDInsight Storm Cluster

A. HDInsight Spark Cluster

You plan to perform batch processing in Azure Databricks once daily.Which type of Databricks cluster should you use? A. job B. interactive C. High Concurrency

A. job

A company has a SaaS solution that uses Azure SQL Database with elastic pools. The solution will have a dedicated database for each customer organization.Customer organizations have peak usage at different periods during the year.Which two factors affect your costs when sizing the Azure SQL Database elastic pools? Each correct answer presents a complete solution.NOTE: Each correct selection is worth one point. A. maximum data size B. number of databases C. eDTUs consumption D. number of read operations E. number of transactions

AC

A company plans to use Azure SQL Database to support a mission-critical application.The application must be highly available without performance degradation during maintenance windows.You need to implement the solution.Which three technologies should you implement? Each correct answer presents part of the solution.NOTE: Each correct selection is worth one point. A. Premium service tier B. Virtual machine Scale Sets C. Basic service tier D. SQL Data Sync E. Always On availability groups F. Zone-redundant configuration

AEF

A company runs Microsoft SQL Server in an on-premises virtual machine (VM).You must migrate the database to Azure SQL Database. You synchronize users from Active Directory to Azure Active Directory (Azure AD).You need to configure Azure SQL Database to use an Azure AD user as administrator.What should you configure? A. For each Azure SQL Database, set the Access Control to administrator. B. For each Azure SQL Database server, set the Active Directory to administrator. C. For each Azure SQL Database, set the Active Directory administrator role. D. For each Azure SQL Database server, set the Access Control to administrator.

B. For each Azure SQL Database server, set the Active Directory to administrator.

Your company uses several Azure HDInsight clusters.The data engineering team reports several errors with some applications using these clusters.You need to recommend a solution to review the health of the clusters.What should you include in your recommendation? A. Azure Automation B. Log Analytics C. Application Insights Reveal Solution

B. Log Analytics

You develop data engineering solutions for a company.You must integrate the company's on-premises Microsoft SQL Server data with Microsoft Azure SQL Database. Data must be transformed incrementally.You need to implement the data integration solution.Which tool should you use to configure a pipeline to copy data? A. Use the Copy Data tool with Blob storage linked service as the source B. Use Azure PowerShell with SQL Server linked service as a source C. Use Azure Data Factory UI with Blob storage linked service as a source D. Use the .NET Data Factory API with Blob storage linked service as the source

B. Use Azure PowerShell with SQL Server linked service as a source

You are developing a data engineering solution for a company. The solution will store a large set of key-value pair data by using Microsoft Azure Cosmos DB.The solution has the following requirements:✑ Data must be partitioned into multiple containers.✑ Data containers must be configured separately.✑ Data must be accessible from applications hosted around the world.✑ The solution must minimize latency.You need to provision Azure Cosmos DB. A. Cosmos account-level throughput. B. Provision an Azure Cosmos DB account with the Azure Table API. Enable geo-redundancy. C. Configure table-level throughput. D. Replicate the data globally by manually adding regions to the Azure Cosmos DB account. E. Provision an Azure Cosmos DB account with the Azure Table API. Enable multi-region writes.

BDE

You manage a solution that uses Azure HDInsight clusters.You need to implement a solution to monitor cluster performance and status.Which technology should you use? A. Azure HDInsight .NET SDK B. Azure HDInsight REST API C. Ambari REST API D. Azure Log Analytics E. Ambari Web UI

C. Ambari REST API

You are developing a solution that will stream to Azure Stream Analytics. The solution will have both streaming data and reference data.Which input type should you use for the reference data? A. Azure Cosmos DB B. Azure Event Hubs C. Azure Blob storage D. Azure IoT Hu

C. Azure Blob storage

The data engineering team manages Azure HDInsight clusters. The team spends a large amount of time creating and destroying clusters daily because most of the data pipeline process runs in minutes.You need to implement a solution that deploys multiple HDInsight clusters with minimal effort.What should you implement? A. Azure Databricks B. Azure Traffic Manager C. Azure Resource Manager templates D. Ambari web user interface

C. Azure Resource Manager templates

You develop data engineering solutions for a company.You must integrate the company's on-premises Microsoft SQL Server data with Microsoft Azure SQL Database. Data must be transformed incrementally.You need to implement the data integration solution.Which tool should you use to configure a pipeline to copy data? A. Use the Copy Data tool with Blob storage linked service as the source B. Use Azure PowerShell with SQL Server linked service as a source C. Use Azure Data Factory UI with Blob storage linked service as a source D. Use the .NET Data Factory API with Blob storage linked service as the source

C. Use Azure Data Factory UI with Blob storage linked service as a source

You use Azure Stream Analytics to receive Twitter data from Azure Event Hubs and to output the data to an Azure Blob storage account.You need to output the count of tweets during the last five minutes every five minutes. Each tweet must only be counted once.Which windowing function should you use? A. a five-minute Session window B. a five-minute Sliding window C. a five-minute Tumbling window D. a five-minute Hopping window that has one-minute hop

C. a five-minute Tumbling window

You plan to create a dimension table in Azure Data Warehouse that will be less than 1 GB.You need to create the table to meet the following requirements:✑ Provide the fastest query time.✑ Minimize data movement.Which type of table should you use? A. hash distributed B. heap C. replicated D. round-robin

C. replicated

A company uses Azure SQL Database to store sales transaction data. Field sales employees need an offline copy of the database that includes last year's sales on their laptops when there is no internet connection available.You need to create the offline export copy.Which three options can you use? Each correct answer presents a complete solution.NOTE: Each correct selection is worth one point. A. Export to a BACPAC file by using Azure Cloud Shell, and save the file to an Azure storage account B. Export to a BACPAC file by using SQL Server Management Studio. Save the file to an Azure storage account C. Export to a BACPAC file by using the Azure portal D. Export to a BACPAC file by using Azure PowerShell and save the file locally E. Export to a BACPAC file by using the SqlPackage utility

CDE

You manage a Microsoft Azure SQL Data Warehouse Gen 2.Users report slow performance when they run commonly used queries. Users do not report performance changes for infrequently used queries.You need to monitor resource utilization to determine the source of the performance issues.Which metric should you monitor? A. Cache used percentage B. Local tempdb percentage C. DWU percentage D. CPU percentage E. Data IO percentage

Correct Answer: A The Gen2 storage architecture automatically tiers your most frequently queried columnstore segments in a cache residing on NVMe based SSDs designed forGen2 data warehouses. Greater performance is realized when your queries retrieve segments that are residing in the cache. You can monitor and troubleshoot slow query performance by determining whether your workload is optimally leveraging the Gen2 cache.Reference:https://docs.microsoft.com/en-us/azure/sql-data-warehouse/sql-data-warehouse-how-to-monitor-cache

A company plans to use Azure Storage for file storage purposes. Compliance rules require:✑ A single storage account to store all operations including reads, writes and deletes✑ Retention of an on-premises copy of historical operationsYou need to configure the storage account.Which two actions should you perform? Each correct answer presents part of the solution.NOTE: Each correct selection is worth one point. A. Configure the storage account to log read, write and delete operations for service type Blob B. Use the AzCopy tool to download log data from $logs/blob C. Configure the storage account to log read, write and delete operations for service-type table D. Use the storage client to download log data from $logs/table E. Configure the storage account to log read, write and delete operations for service type queue Reveal Solution

Correct Answer: ABStorage Logging logs request data in a set of blobs in a blob container named $logs in your storage account. This container does not show up if you list all the blob containers in your account but you can see its contents if you access it directly.To view and analyze your log data, you should download the blobs that contain the log data you are interested in to a local machine. Many storage-browsing tools enable you to download blobs from your storage account; you can also use the Azure Storage team provided command-line Azure Copy Tool (AzCopy) to download your log data.References:https://docs.microsoft.com/en-us/rest/api/storageservices/enabling-storage-logging-and-accessing-log-data

You manage a process that performs analysis of daily web traffic logs on an HDInsight cluster. Each of the 250 web servers generates approximately 10 megabytes (MB) of log data each day. All log data is stored in a single folder in Microsoft Azure Data Lake Storage Gen 2.You need to improve the performance of the process.Which two changes should you make? Each correct answer presents a complete solution.NOTE: Each correct selection is worth one point. A. Combine the daily log files for all servers into one file B. Increase the value of the mapreduce.map.memory parameter C. Move the log files into folders so that each day's logs are in their own folder D. Increase the number of worker nodes E. Increase the value of the hive.tez.container.size parameter

Correct Answer: ACA: Typically, analytics engines such as HDInsight and Azure Data Lake Analytics have a per-file overhead. If you store your data as many small files, this can negatively affect performance. In general, organize your data into larger sized files for better performance (256MB to 100GB in size). Some engines and applications might have trouble efficiently processing files that are greater than 100GB in size.C: For Hive workloads, partition pruning of time-series data can help some queries read only a subset of the data which improves performance.Those pipelines that ingest time-series data, often place their files with a very structured naming for files and folders. Below is a very common example we see for data that is structured by date:\DataSet\YYYY\MM\DD\datafile_YYYY_MM_DD.tsvNotice that the datetime information appears both as folders and in the filename.References:https://docs.microsoft.com/en-us/azure/storage/blobs/data-lake-storage-performance-tuning-guidance

You develop data engineering solutions for a company.A project requires the deployment of data to Azure Data Lake Storage.You need to implement role-based access control (RBAC) so that project members can manage the Azure Data Lake Storage resources.Which three actions should you perform? Each correct answer presents part of the solution.NOTE: Each correct selection is worth one point. A. Assign Azure AD security groups to Azure Data Lake Storage. B. Configure end-user authentication for the Azure Data Lake Storage account. C. Configure service-to-service authentication for the Azure Data Lake Storage account. D. Create security groups in Azure Active Directory (Azure AD) and add project members. E. Configure access control lists (ACL) for the Azure Data Lake Storage account.

Correct Answer: ADEAD: Create security groups in Azure Active Directory. Assign users or security groups to Data Lake Storage Gen1 accounts.E: Assign users or security groups as ACLs to the Data Lake Storage Gen1 file systemReferences:https://docs.microsoft.com/en-us/azure/data-lake-store/data-lake-store-secure-data

You are monitoring an Azure Stream Analytics job.You discover that the Backlogged Input Events metric is increasing slowly and is consistently non-zero.You need to ensure that the job can handle all the events.What should you do? A. Change the compatibility level of the Stream Analytics job. B. Increase the number of streaming units (SUs). C. Create an additional output stream for the existing input stream. D. Remove any named consumer groups from the connection and use $default.

Correct Answer: B Backlogged Input Events: Number of input events that are backlogged. A non-zero value for this metric implies that your job isn't able to keep up with the number of incoming events. If this value is slowly increasing or consistently non-zero, you should scale out your job. You should increase the Streaming Units.Note: Streaming Units (SUs) represents the computing resources that are allocated to execute a Stream Analytics job. The higher the number of SUs, the moreCPU and memory resources are allocated for your job.Reference:https://docs.microsoft.com/bs-cyrl-ba/azure/stream-analytics/stream-analytics-monitoring

You have an Azure data solution that contains an Azure SQL data warehouse named DW1.Several users execute adhoc queries to DW1 concurrently.You regularly perform automated data loads to DW1.You need to ensure that the automated data loads have enough memory available to complete quickly and successfully when the adhoc queries runWhat should you do? A. Hash distribute the large fact tables in DW1 before performing the automated data loads. B. Assign a larger resource class to the automated data load queries. C. Create sampled statistics for every column in each table of DW1. D. Assign a smaller resource class to the automated data load queries.

Correct Answer: B To ensure the loading user has enough memory to achieve maximum compression rates, use loading users that are a member of a medium or large resource class.References:https://docs.microsoft.com/en-us/azure/sql-data-warehouse/guidance-for-loading-data

You have an Azure data solution that contains an Azure SQL data warehouse named DW1.Several users execute adhoc queries to DW1 concurrently.You regularly perform automated data loads to DW1.You need to ensure that the automated data loads have enough memory available to complete quickly and successfully when the adhoc queries runWhat should you do? A. Hash distribute the large fact tables in DW1 before performing the automated data loads. B. Assign a larger resource class to the automated data load queries. C. Create sampled statistics for every column in each table of DW1. D. Assign a smaller resource class to the automated data load queries.

Correct Answer: B To ensure the loading user has enough memory to achieve maximum compression rates, use loading users that are a member of a medium or large resource class.References:https://docs.microsoft.com/en-us/azure/sql-data-warehouse/guidance-for-loading-data

You are monitoring an Azure Stream Analytics job.You discover that the Backlogged Input Events metric is increasing slowly and is consistently non-zero.You need to ensure that the job can handle all the events.What should you do? A. Change the compatibility level of the Stream Analytics job. B. Increase the number of streaming units (SUs). C. Create an additional output stream for the existing input stream. D. Remove any named consumer groups from the connection and use $default.

Correct Answer: B Backlogged Input Events: Number of input events that are backlogged. A non-zero value for this metric implies that your job isn't able to keep up with the number of incoming events. If this value is slowly increasing or consistently non-zero, you should scale out your job. You should increase the Streaming Units.Note: Streaming Units (SUs) represents the computing resources that are allocated to execute a Stream Analytics job. The higher the number of SUs, the moreCPU and memory resources are allocated for your job.Reference:https://docs.microsoft.com/bs-cyrl-ba/azure/stream-analytics/stream-analytics-monitoring

You have an Azure Storage account that contains 100 GB of files. The files contain text and numerical values. 75% of the rows contain description data that has an average length of 1.1 MB.You plan to copy the data from the storage account to an Azure SQL data warehouse.You need to prepare the files to ensure that the data copies quickly.Solution: You copy the files to a table that has a columnstore index.Does this meet the goal? A. Yes B. No

Correct Answer: B Instead modify the files to ensure that each row is less than 1 MB.References:https://docs.microsoft.com/en-us/azure/sql-data-warehouse/guidance-for-loading-data

A project requires the deployment of resources to Microsoft Azure for batch data processing on Azure HDInsight. Batch processing will run daily and must:✑ Scale to minimize costs✑ Be monitored for cluster performanceYou need to recommend a tool that will monitor clusters and provide information to suggest how to scale.Solution: Monitor cluster load using the Ambari Web UI.Does the solution meet the goal?

Correct Answer: BAmbari Web UI does not provide information to suggest how to scale.Instead monitor clusters by using Azure Log Analytics and HDInsight cluster management solutions.References:https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-oms-log-analytics-tutorial https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-manage-ambari

You develop data engineering solutions for a company.You need to ingest and visualize real-time Twitter data by using Microsoft Azure.Which three technologies should you use? Each correct answer presents part of the solution.NOTE: Each correct selection is worth one point. A. Event Grid topic B. Azure Stream Analytics Job that queries Twitter data from an Event Hub C. Azure Stream Analytics Job that queries Twitter data from an Event Grid D. Logic App that sends Twitter posts which have target keywords to Azure E. Event Grid subscription F. Event Hub instance

Correct Answer: BDF You can use Azure Logic apps to send tweets to an event hub and then use a Stream Analytics job to read from event hub and send them to PowerBI.

An application will use Microsoft Azure Cosmos DB as its data solution. The application will use the Cassandra API to support a column-based database type that uses containers to store items.You need to provision Azure Cosmos DB. Which container name and item name should you use? Each correct answer presents part of the solutions.NOTE: Each correct answer selection is worth one point. A. collection B. rows C. graph D. entities E. table

Correct Answer: BE

A company has a real-time data analysis solution that is hosted on Microsoft Azure. The solution uses Azure Event Hub to ingest data and an Azure StreamAnalytics cloud job to analyze the data. The cloud job is configured to use 120 Streaming Units (SU).You need to optimize performance for the Azure Stream Analytics job.Which two actions should you perform? Each correct answer presents part of the solution.NOTE: Each correct selection is worth one point. A. Implement event ordering B. Scale the SU count for the job up C. Implement Azure Stream Analytics user-defined functions (UDF) D. Scale the SU count for the job down E. Implement query parallelization by partitioning the data output F. Implement query parallelization by partitioning the data input

Correct Answer: BFScale out the query by allowing the system to process each input partition separately.F: A Stream Analytics job definition includes inputs, a query, and output. Inputs are where the job reads the data stream from.References:https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-parallelization

A project requires the deployment of resources to Microsoft Azure for batch data processing on Azure HDInsight. Batch processing will run daily and must:✑ Scale to minimize costsBe monitored for cluster performanceYou need to recommend a tool that will monitor clusters and provide information to suggest how to scale.Solution: Download Azure HDInsight cluster logs by using Azure PowerShell.Does the solution meet the goal?

Correct Answer: BInstead monitor clusters by using Azure Log Analytics and HDInsight cluster management solutions.References:https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-oms-log-analytics-tutorial

You have an Azure Stream Analytics query. The query returns a result set that contains 10,000 distinct values for a column named clusterID.You monitor the Stream Analytics job and discover high latency.You need to reduce the latency.Which two actions should you perform? Each correct answer presents a complete solution.NOTE: Each correct selection is worth one point. A. Add a pass-through query. B. Add a temporal analytic function. C. Scale out the query by using PARTITION BY. D. Convert the query to a reference query. E. Increase the number of streaming units.

Correct Answer: CE C: Scaling a Stream Analytics job takes advantage of partitions in the input or output. Partitioning lets you divide data into subsets based on a partition key. A process that consumes the data (such as a Streaming Analytics job) can consume and write different partitions in parallel, which increases throughput.E: Streaming Units (SUs) represents the computing resources that are allocated to execute a Stream Analytics job. The higher the number of SUs, the more CPU and memory resources are allocated for your job. This capacity lets you focus on the query logic and abstracts the need to manage the hardware to run yourStream Analytics job in a timely manner.References:https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-parallelization https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-streaming-unit-consumption

You have an Azure SQL database that has masked columns.You need to identify when a user attempts to infer data from the masked columns.What should you use? A. Azure Advanced Threat Protection (ATP) B. custom masking rules C. Transparent Data Encryption (TDE) D. auditing

Correct Answer: DDynamic Data Masking is designed to simplify application development by limiting data exposure in a set of pre-defined queries used by the application. WhileDynamic Data Masking can also be useful to prevent accidental exposure of sensitive data when accessing a production database directly, it is important to note that unprivileged users with ad-hoc query permissions can apply techniques to gain access to the actual data. If there is a need to grant such ad-hoc access,Auditing should be used to monitor all database activity and mitigate this scenario.References:https://docs.microsoft.com/en-us/sql/relational-databases/security/dynamic-data-masking

You are a data engineer implementing a lambda architecture on Microsoft Azure. You use an open-source big data solution to collect, process, and maintain data.The analytical data store performs poorly.You must implement a solution that meets the following requirements:✑ Provide data warehousing✑ Reduce ongoing management activities✑ Deliver SQL query responses in less than one secondYou need to create an HDInsight cluster to meet the requirements.Which type of cluster should you create? A. Interactive Query B. Apache Hadoop C. Apache HBase D. Apache Spark

D

A company manages several on-premises Microsoft SQL Server databases.You need to migrate the databases to Microsoft Azure by using a backup process of Microsoft SQL Server.Which data technology should you use? A. Azure SQL Database single database B. Azure SQL Data Warehouse C. Azure Cosmos DB D. Azure SQL Database Managed Instance

D. Azure SQL Database Managed Instance Hide Solution

You develop data engineering solutions for a company. The company has on-premises Microsoft SQL Server databases at multiple locations.The company must integrate data with Microsoft Power BI and Microsoft Azure Logic Apps. The solution must avoid single points of failure during connection and transfer to the cloud. The solution must also minimize latency.You need to secure the transfer of data between on-premises databases and Microsoft Azure.What should you do? A. Install a standalone on-premises Azure data gateway at each location B. Install an on-premises data gateway in personal mode at each location C. Install an Azure on-premises data gateway at the primary location D. Install an Azure on-premises data gateway as a cluster at each location

D. Install an Azure on-premises data gateway as a cluster at each location Reveal Solution

You plan to implement an Azure Cosmos DB database that will write 100,000 JSON every 24 hours. The database will be replicated to three regions. Only one region will be writable.You need to select a consistency level for the database to meet the following requirements:✑ Guarantee monotonic reads and writes within a session.✑ Provide the fastest throughput.✑ Provide the lowest latency.Which consistency level should you select? A. Strong B. Bounded Staleness C. Eventual D. Session E. Consistent Prefix

D. Session

You plan to use Microsoft Azure SQL Database instances with strict user access control. A user object must:✑ Move with the database if it is run elsewhere✑ Be able to create additional usersYou need to create the user object with correct permissions.Which two Transact-SQL commands should you run? Each correct answer presents part of the solution.NOTE: Each correct selection is worth one point. A. ALTER LOGIN Mary WITH PASSWORD = 'strong_password'; B. CREATE LOGIN Mary WITH PASSWORD = 'strong_password'; C. ALTER ROLE db_owner ADD MEMBER Mary; D. CREATE USER Mary WITH PASSWORD = 'strong_password'; E. GRANT ALTER ANY USER TO Mary;

DE

A company is designing a hybrid solution to synchronize data and on-premises Microsoft SQL Server database to Azure SQL Database.You must perform an assessment of databases to determine whether data will move without compatibility issues. You need to perform the assessment.Which tool should you use? A. SQL Server Migration Assistant (SSMA) B. Microsoft Assessment and Planning Toolkit C. SQL Vulnerability Assessment (VA) D. Azure SQL Data Sync E. Data Migration Assistant (DMA)

E

You are a data architect. The data engineering team needs to configure a synchronization of data between an on-premises Microsoft SQL Server database toAzure SQL Database.Ad-hoc and reporting queries are being overutilized the on-premises production instance. The synchronization process must:✑ Perform an initial data synchronization to Azure SQL Database with minimal downtime✑ Perform bi-directional data synchronization after initial synchronizationYou need to implement this synchronization solution.Which synchronization method should you use? A. transactional replication B. Data Migration Assistant (DMA) C. backup and restore D. SQL Server Agent job E. Azure SQL Data Sync

E. Azure SQL Data Sync Hide Solution

Each day, company plans to store hundreds of files in Azure Blob Storage and Azure Data Lake Storage. The company uses the parquet format.You must develop a pipeline that meets the following requirements:✑ Process data every six hours✑ Offer interactive data analysis capabilities✑ Offer the ability to process data using solid-state drive (SSD) caching✑ Use Directed Acyclic Graph(DAG) processing mechanisms✑ Provide support for REST API calls to monitor processes✑ Provide native support for Python✑ Integrate with Microsoft Power BIYou need to select the appropriate data technology to implement the pipeline.Which data technology should you implement? A. Azure SQL Data Warehouse B. HDInsight Apache Storm cluster C. Azure Stream Analytics D. HDInsight Apache Hadoop cluster using MapReduce E. HDInsight Spark cluster

E. HDInsight Spark cluster

A company has a SaaS solution that uses Azure SQL Database with elastic pools. The solution contains a dedicated database for each customer organization.Customer organizations have peak usage at different periods during the year.You need to implement the Azure SQL Database elastic pool to minimize cost.Which option or options should you configure? A. Number of transactions only B. eDTUs per database only C. Number of databases only D. CPU usage only E. eDTUs and max data size

E. eDTUs and max data size Hide Solution

A company has a Microsoft Azure HDInsight solution that uses different cluster types to process and analyze data. Operations are continuous.Reports indicate slowdowns during a specific time window.You need to determine a monitoring solution to track down the issue in the least amount of time.What should you use? A. Azure Log Analytics log search query B. Ambari REST API C. Azure Monitor Metrics D. HDInsight .NET SDK E. Azure Log Analytics alert rule query

c

You configure monitoring for a Microsoft Azure SQL Data Warehouse implementation. The implementation uses PolyBase to load data from comma-separated value (CSV) files stored in Azure Data Lake Gen 2 using an external table.Files with an invalid schema cause errors to occur.You need to monitor for an invalid schema error.For which error should you monitor? A. EXTERNAL TABLE access failed due to internal error: 'Java exception raised on call to HdfsBridge_Connect: Error [com.microsoft.polybase.client.KerberosSecureLogin] occurred while accessing external file.' B. EXTERNAL TABLE access failed due to internal error: 'Java exception raised on call to HdfsBridge_Connect: Error [No FileSystem for scheme: wasbs] occurred while accessing external file.' C. Cannot execute the query "Remote Query" against OLE DB provider "SQLNCLI11": for linked server "(null)", Query aborted- the maximum reject threshold (o rows) was reached while reading from an external source: 1 rows rejected out of total 1 rows processed. D. EXTERNAL TABLE access failed due to internal error: 'Java exception raised on call to HdfsBridge_Connect: Error [Unable to instantiate LoginClass] occurred while accessing external file.' Reveal Solution Discussion 2 Question #8

c

You are the data engineer for your company. An application uses a NoSQL database to store data. The database uses the key-value and wide-column NoSQL database type.Developers need to access data in the database using an API.You need to determine which API to use for the database model and type.Which two APIs should you use? Each correct answer presents a complete solution.NOTE: Each correct selection is worth one point. A. Table API B. MongoDB API C. Gremlin API D. SQL API E. Cassandra API

key value --> Table API graph --> Gremlin API document --> SQL API and MongoDB API columnar --> Cassandra API so, the answer is A,E


Ensembles d'études connexes

3.8 Off-Balance Sheet Activities

View Set

Stats Quiz 9 Measure of Central Tendency

View Set