DRAG DROP
Overview:
Relecloud is a social media company that processes hundreds of millions of social media posts per day and
sells advertisements to several hundred companies.
Relecloud has a Microsoft SQL Server database named DB1 that stores information about the advertisers. DB1
is hosted on a Microsoft Azure virtual machine.
Relecloud has two main offices. The offices are located in San Francisco and New York City.
The offices connect to each other by using a site-to-site VPN. Each office connects directly to the Internet.
Relecloud modifies the pricing of its advertisements based on trending topics. Topics are considered to be
trending if they generate many mentions in a specific country during a 15-minute time frame. The highest
trending topics generate the highest advertising revenue.
Relecloud wants to deliver reports to the advertisers by using Microsoft Power BI. The reports will provide realtime data on trending topics, current advertising rates, and advertising costs for a given month.
Relecloud will analyze the trending topics data, and then store the data in a new data warehouse for ad-hoc
analysis. The data warehouse is expected to grow at a rate of 1 GB per hour or 8.7 terabytes (TB) per year.
The data will be retained for five years for the purpose of long-term trending.
Requirements:
Management at Relecloud must be able to view which topics are trending to adjust advertising rates in near
real-time.
Relecloud plans to implement a new streaming analytics platform that will report on trending topics.
Relecloud plans to implement a data warehouse named DB2.
Relecloud identifies the following technical requirements:
Social media data must be analyzed to identify trending topics in real-time.
The use of Infrastructure as a Service (IaaS) platforms must minimized, whenever possible.
The real-time solution used to analyze the social media data must support scaling up and down without
service interruption.
Relecloud identifies the following technical requirements for the advertisers:
The advertisers must be able to see only their own data in the Power BI reports.
The advertisers must authenticate to Power BI by using Azure Active Directory (Azure AD) credentials.
The advertisers must be able to leverage existing Transact-SQL language knowledge when developing the
real-time streaming solution.
Members of the internal advertising sales team at Relecloud must be able to see only the sales date of the
advertisers to which they are assigned.
The internal Relecloud advertising sales team must be prevented from inserting, updating, and deleting
rows for the advertisers to which they are not assigned.
The internal Relecloud advertising sales team must be able to use a text file to update the list of advertisers,
and then to upload the file to Azure Blob storage.
Relecloud identifies the following requirements for DB1:
Data generated by the streaming analytics platform must be stored in DB1.
The user names of the advertisers must be mapped to CustomerID in a table named Table2.The advertisers in DB1 must be stored in a table named Table1 and must be refreshed nightly.
The user names of the employees at Relecloud must be mapped to EmployeeID in a table named Table3.
Relecloud identifies the following requirements for DB2:
DB2 must have minimal storage costs.
DB2 must run load processes in parallel.
DB2 must support massive parallel processing.
DB2 must be able to store more than 40 TB of data.
DB2 must support scaling up and down, as required.
Data from DB1 must be archived in DB2 for long-term storage.
All of the reports that are executed from DB2 must use aggregation.
Users must be able to pause DB2 when the data warehouse is not in use.
Users must be able to view previous versions of the data in DB2 by using aggregates.
Relecloud identifies the following requirements for extract, transformation, and load (ETL):
Data movement between DB1 and DB2 must occur each hour.
An email alert must be generated when a failure of any type occurs during ETL processing.
Sample code and data:
You execute the following code for a table named rls_table1.
You use the following code to create Table1.
create table table1
(customerid int,
salespersonid int
…
)
Go
The following is a sample of the streaming data.
You need to create a query that identifies the trending topics.
How should you complete the query? To answer, drag the appropriate values to the correct targets. Each value
may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to
view content.
NOTE: Each correct selection is worth one point.
Select and Place:
Explanation:
From scenario: Topics are considered to be trending if they generate many mentions in a specific country
during a 15-minute time frame.
Box 1: TimeStamp
Azure Stream Analytics (ASA) is a cloud service that enables real-time processing over streams of data flowing
in from devices, sensors, websites and other live systems. The stream-processing logic in ASA is expressed in
a SQL-like query language with some added extensions such as windowing for performing temporal
calculations.
ASA is a temporal system, so every event that flows through it has a timestamp. A timestamp is assigned
automatically based on the event’s arrival time to the input source but you can also access a timestamp in your
event payload explicitly using TIMESTAMP BY:
SELECT * FROM SensorReadings TIMESTAMP BY time
Box 2: GROUP BY
Example: Generate an output event if the temperature is above 75 for a total of 5 seconds
SELECT sensorId, MIN(temp) as temp
FROM SensorReadings
TIMESTAMP BY time
GROUP BY sensorId, SlidingWindow(second, 5)
HAVING MIN(temp) > 75
Box 3: SlidingWindow
Windowing is a core requirement for stream processing applications to perform set-based operations like
counts or aggregations over events that arrive within a specified period of time. ASA supports three types of
windows: Tumbling, Hopping, and Sliding.With a Sliding Window, the system is asked to logically consider all possible windows of a given length and
output events for cases when the content of the window actually changes – that is, when an event entered or
existed the window.
https://blogs.technet.microsoft.com/machinelearning/2015/06/01/the-azure-stream-analytics-querylanguage/