Which function should you use for the time-series data …
You have sensor devices that report data to Microsoft Azure Stream Analytics. Each sensor reports data
several times per second.
You need to create a live dashboard in Microsoft Power BI that shows the performance of the sensor devices.
The solution must minimize lag when visualizing the data.
Which function should you use for the time-series data element?
Which three actions should you perform in sequence to m…
DRAG DROP
Note: This question is part of a series of questions that present the same scenario. For your convenience, the
scenario is repeated in each question. Each question presents a different goal and answer choices, but the text
of the scenario is exactly the same in each question in this series.
Start of repeated scenario
You are migrating an existing on-premises data warehouse named LocalDW to Microsoft Azure. You will use
an Azure SQL data warehouse named AzureDW for data storage and an Azure Data Factory named AzureDF
for extract, transformation, and load (ETL) functions.
For each table in LocalDW, you create a table in AzureDW.
On the on-premises network, you have a Data Management Gateway.
Some source data is stored in Azure Blob storage. Some source data is stored on an on-premises Microsoft
SQL Server instance. The instance has a table named Table1.
After data is processed by using AzureDF, the data must be archived and accessible forever. The archived data
must meet a Service Level Agreement (SLA) for availability of 99 percent. If an Azure region fails, the archived
data must be available for reading always. The storage solution for the archived data must minimize costs.
End of repeated scenario.
Which three actions should you perform in sequence to migrate the on-premises data warehouse to Azure SQL
Data Warehouse? To answer, move the appropriate actions from the list of actions to the answer area and
arrange them in the correct order.
Select and Place:
Which Window function should you use for each requirement?
DRAG DROP
You plan to develop a solution for real-time sentiment analysis of Twitter data.
You need to create a Microsoft Azure Stream Analytics job query to count the number of tweets during a period.
Which Window function should you use for each requirement? To answer, drag the appropriate functions to the
correct requirements. Each function may be used once, more than once, or not at all. You may need to drag thesplit bar between panes or scroll to view content.
Select and Place:
How should you complete the statement?
DRAG DROP
You use Microsoft Azure Stream Analytics to analyze data from an Azure event hub in real time and send the
output to a table named Table1 in an Azure SQL database. Table1 has three columns named Date, EventID,
and User.
You need to prevent duplicate data from being stored in the database.
How should you complete the statement? To answer, drag the appropriate values to the correct targets. Each
value may be used once, more than once, or not at all. You may need to drag the split bar between panes or
scroll to view content.
NOTE: Each correct selection is worth one point.
Select and Place:
How should you complete the Stream Analytics query?
DRAG DROP
You have IoT devices that produce the following output.
You need to use Microsoft Azure Stream Analytics to convert the output into the tabular format described in the
following table.
How should you complete the Stream Analytics query? To answer, drag the appropriate values to the correct
targets. Each value may be used once, more than once, or not at all. You may need to drag the split bar
between panes or scroll to view content.
NOTE: Each correct selection is worth one point.
Select and Place:
How should you configure the storage to archive the sou…
HOTSPOT
Note: This question is part of a series of questions that present the same scenario. For your convenience, the
scenario is repeated in each question. Each question presents a different goal and answer choices, but the text
of the scenario is exactly the same in each question in this series.Start of repeated scenario
You are migrating an existing on-premises data warehouse named LocalDW to Microsoft Azure. You will use
an Azure SQL data warehouse named AzureDW for data storage and an Azure Data Factory named AzureDF
for extract, transformation, and load (ETL) functions.
For each table in LocalDW, you create a table in AzureDW.
On the on-premises network, you have a Data Management Gateway.
Some source data is stored in Azure Blob storage. Some source data is stored on an on-premises Microsoft
SQL Server instance. The instance has a table named Table1.
After data is processed by using AzureDF, the data must be archived and accessible forever. The archived data
must meet a Service Level Agreement (SLA) for availability of 99 percent. If an Azure region fails, the archived
data must be available for reading always. The storage solution for the archived data must minimize costs.
End of repeated scenario.
How should you configure the storage to archive the source data? To answer, select the appropriate options in
the answer area.
NOTE: Each correct selection is worth one point.
Hot Area:
What should you create?
Note: This question is part of a series of questions that present the same scenario. For your convenience, the
scenario is repeated in each question. Each question presents a different goal and answer choices, but the text
of the scenario is exactly the same in each question in this series.
Start of repeated scenario
You are migrating an existing on-premises data warehouse named LocalDW to Microsoft Azure. You will use
an Azure SQL data warehouse named AzureDW for data storage and an Azure Data Factory named AzureDF
for extract, transformation, and load (ETL) functions.
For each table in LocalDW, you create a table in AzureDW.
On the on-premises network, you have a Data Management Gateway.
Some source data is stored in Azure Blob storage. Some source data is stored on an on-premises Microsoft
SQL Server instance. The instance has a table named Table1.
After data is processed by using AzureDF, the data must be archived and accessible forever. The archived data
must meet a Service Level Agreement (SLA) for availability of 99 percent. If an Azure region fails, the archived
data must be available for reading always. The storage solution for the archived data must minimize costs.
End of repeated scenario.
You need to define the schema of Table1 in AzureDF.
What should you create?
You need to configure Azure Data Factory to connect to …
Note: This question is part of a series of questions that present the same scenario. For your convenience, the
scenario is repeated in each question. Each question presents a different goal and answer choices, but the text
of the scenario is exactly the same in each question in this series.
Start of repeated scenario
You are migrating an existing on-premises data warehouse named LocalDW to Microsoft Azure. You will use
an Azure SQL data warehouse named AzureDW for data storage and an Azure Data Factory named AzureDF
for extract, transformation, and load (ETL) functions.
For each table in LocalDW, you create a table in AzureDW.
On the on-premises network, you have a Data Management Gateway.
Some source data is stored in Azure Blob storage. Some source data is stored on an on-premises Microsoft
SQL Server instance. The instance has a table named Table1.
After data is processed by using AzureDF, the data must be archived and accessible forever. The archived data
must meet a Service Level Agreement (SLA) for availability of 99 percent. If an Azure region fails, the archived
data must be available for reading always.
End of repeated scenario.
You need to configure Azure Data Factory to connect to the on-premises SQL Server instance.
What should you do first?
What should you create?
Note: This question is part of a series of questions that present the same scenario. For your convenience, the
scenario is repeated in each question. Each question presents a different goal and answer choices, but the text
of the scenario is exactly the same in each question in this series.
Start of repeated scenario
You are migrating an existing on-premises data warehouse named LocalDW to Microsoft Azure. You will use
an Azure SQL data warehouse named AzureDW for data storage and an Azure Data Factory named AzureDF
for extract, transformation, and load (ETL) functions.
For each table in LocalDW, you create a table in AzureDW.On the on-premises network, you have a Data Management Gateway.
Some source data is stored in Azure Blob storage. Some source data is stored on an on-premises Microsoft
SQL Server instance. The instance has a table named Table1.
After data is processed by using AzureDF, the data must be archived and accessible forever. The archived data
must meet a Service Level Agreement (SLA) for availability of 99 percent. If an Azure region fails, the archived
data must be available for reading always.
End of repeated scenario.
You need to connect AzureDF to the storage account.
What should you create?
What should you create?
Note: This question is part of a series of questions that present the same scenario. For your convenience, the
scenario is repeated in each question. Each question presents a different goal and answer choices, but the text
of the scenario is exactly the same in each question in this series.
Start of repeated scenario
You are migrating an existing on-premises data warehouse named LocalDW to Microsoft Azure. You will use
an Azure SQL data warehouse named AzureDW for data storage and an Azure Data Factory named AzureDF
for extract, transformation, and load (ETL) functions.
For each table in LocalDW, you create a table in AzureDW.
On the on-premises network, you have a Data Management Gateway.
Some source data is stored in Azure Blob storage. Some source data is stored on an on-premises Microsoft
SQL Server instance. The instance has a table named Table1.
After data is processed by using AzureDF, the data must be archived and accessible forever. The archived data
must meet a Service Level Agreement (SLA) for availability of 99 percent. If an Azure region fails, the archived
data must be available for reading always.
End of repeated scenario.
You need to configure an activity to move data from blob storage to AzureDW.
What should you create?