You plan to use Microsoft Azure Event Hubs to ingest sensor data. You plan to use Azure Stream Analytics to analyze the data in real time and to send the output
directly to Azure Data Lake Store.
You need to write events to the Data Lake Store in batches.
What should you use?
You are developing an application that uses Microsoft Azure Stream Analytics.
You have data structures that are defined dynamically.
You want to enable consistency between the logical methods used by stream processing and batch processing.
You need to ensure that the data can be integrated by using consistent data points.
What should you use to process the data?
Note: This question is part of series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
You are monitoring user queries to a Microsoft Azure SQL data warehouse that has six compute nodes.
You discover that compute node utilization is uneven. The rows_processed column from sys.dm_pdw_dms_workers shows a significant variation in the number of rows being moved among the distributions. The variation occurs for the same table, using the same query.
You need to ensure that the load is distributed evenly across the compute nodes.
Solution: You add a clustered columnstore index.
Does this meet the goal?
Note: This question is part of series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
You are monitoring user queries to a Microsoft Azure SQL data warehouse that has six compute nodes.
You discover that compute node utilization is uneven. The rows_processed column from sys.dm_pdw_dms_workers shows a significant variation in the number of rows being moved among the distributions. The variation occurs for the same table, using the same query.
You need to ensure that the load is distributed evenly across the compute nodes.
Solution: You configure joined tables to use a joined distribution column.
Does this meet the goal?
Note: This question is part of series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution.
After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
You are monitoring user queries to a Microsoft Azure SQL data warehouse that has six compute nodes.
You discover that compute node utilization is uneven. The rows_processed column from sys.dm_pdw_dms_workers shows a significant variation in the number of rows being moved among the distributions. The variation occurs for the same table, using the same query.
You need to ensure that the load is distributed evenly across the compute nodes.
Solution: You add a nonclustered columnstore index.
Does this meet the goal?
You use Microsoft Azure Data Lake Store as the default storage for an Azure HDInsight cluster.
You establish an SSH connection to the HDInsight cluster.
You need to copy files from the HDInsight cluster to the Data Lake Store.
Which command should you use?
You are designing a solution that will use Microsoft Azure Data Lake Store.
You need to recommend a solution to ensure that the storage service is available if a regional outage occurs. The solution must minimize costs.
What should you recommend?
You have a Microsoft Azure Data Lake Analytics service and an Azure Data Lake Store.
You need to use Python to submit a U-SQL job.
Which Python module should you install?
You are implementing a solution by using Microsoft Azure Data Lake Analytics.
You have a dataset that contains data-related to website visits.
You need to combine overlapping visits into a single entry based on the timestamp of the visits.
Which type of U-SQL interface should you use?
You plan to add a file from Microsoft Azure Data Lake Store to Azure Data Catalog.
You run the Data Catalog tool and select Data Lake Store as the data source.
Which information should you enter in the Store Account field to connect to the Data Lake Store?