Microsoft Exam

Similar documents
exam. Microsoft Perform Data Engineering on Microsoft Azure HDInsight. Version 1.0

Microsoft Perform Data Engineering on Microsoft Azure HDInsight.

Exam Questions

Microsoft. Exam Questions Perform Data Engineering on Microsoft Azure HDInsight (beta) Version:Demo

microsoft

Stages of Data Processing

Microsoft. Exam Questions Perform Data Engineering on Microsoft Azure HDInsight (beta) Version:Demo

Flash Storage Complementing a Data Lake for Real-Time Insight

BIG DATA COURSE CONTENT

Big Data Technology Ecosystem. Mark Burnette Pentaho Director Sales Engineering, Hitachi Vantara

Modern Data Warehouse The New Approach to Azure BI

Microsoft Analyzing and Visualizing Data with Power BI.

Big Data com Hadoop. VIII Sessão - SQL Bahia. Impala, Hive e Spark. Diógenes Pires 03/03/2018

Processing Unstructured Data. Dinesh Priyankara Founder/Principal Architect dinesql Pvt Ltd.

Microsoft Azure Databricks for data engineering. Building production data pipelines with Apache Spark in the cloud

Topics. Big Data Analytics What is and Why Hadoop? Comparison to other technologies Hadoop architecture Hadoop ecosystem Hadoop usage examples

HDInsight > Hadoop. October 12, 2017

From Single Purpose to Multi Purpose Data Lakes. Thomas Niewel Technical Sales Director DACH Denodo Technologies March, 2019

Overview of Data Services and Streaming Data Solution with Azure

Vendor: Microsoft. Exam Code: Exam Name: Developing SQL Data Models. Version: Demo

@Pentaho #BigDataWebSeries

Microsoft Design and Implement Cloud Data Platform Solutions. Download Full version :

Microsoft Analytics Platform System (APS)

Big Data with Hadoop Ecosystem

Microsoft. Perform Data Engineering on Microsoft Azure HDInsight Version: Demo. Web: [ Total Questions: 10]

White Paper / Azure Data Platform: Ingest

Gain Insights From Unstructured Data Using Pivotal HD. Copyright 2013 EMC Corporation. All rights reserved.

MODERN BIG DATA DESIGN PATTERNS CASE DRIVEN DESINGS

Security and Performance advances with Oracle Big Data SQL

SQT03 Big Data and Hadoop with Azure HDInsight Andrew Brust. Senior Director, Technical Product Marketing and Evangelism

Approaching the Petabyte Analytic Database: What I learned

Modernizing Business Intelligence and Analytics

Acquiring Big Data to Realize Business Value

Lambda Architecture for Batch and Stream Processing. October 2018

Asanka Padmakumara. ETL 2.0: Data Engineering with Azure Databricks

Lambda Architecture for Batch and Real- Time Processing on AWS with Spark Streaming and Spark SQL. May 2015

Data Architectures in Azure for Analytics & Big Data

Big Data Architect.

Microsoft Exam Questions and Answers (PDF) Microsoft Exam Questions BrainDumps

Enable IoT Solutions using Azure

BraindumpsQA. IT Exam Study materials / Braindumps

Big Data on AWS. Big Data Agility and Performance Delivered in the Cloud. 2015, Amazon Web Services, Inc. or its Affiliates. All rights reserved.

Data sources. Gartner, The State of Data Warehousing in 2012

Prototyping Data Intensive Apps: TrendingTopics.org

Increase Value from Big Data with Real-Time Data Integration and Streaming Analytics

Streaming Integration and Intelligence For Automating Time Sensitive Events

exam.105q Microsoft Implementing Data Models and Reports with Microsoft SQL Server 2014

Cloud has become the New Normal

Building a Data Strategy for a Digital World

Exam Questions

Blended Learning Outline: Developer Training for Apache Spark and Hadoop (180404a)

5/24/ MVP SQL Server: Architecture since 2010 MCT since 2001 Consultant and trainer since 1992

Data sources. Gartner, The State of Data Warehousing in 2012

BEST BIG DATA CERTIFICATIONS

How to Troubleshoot Databases and Exadata Using Oracle Log Analytics

Built for Speed: Comparing Panoply and Amazon Redshift Rendering Performance Utilizing Tableau Visualizations

Abstract. The Challenges. ESG Lab Review InterSystems IRIS Data Platform: A Unified, Efficient Data Platform for Fast Business Insight

EXAM PRO:MS SQL 2008, Designing a Business Intelligence. Buy Full Product.

Microsoft Big Data and Hadoop

Přehled novinek v SQL Server 2016

Tuning Enterprise Information Catalog Performance

Big Data. Big Data Analyst. Big Data Engineer. Big Data Architect

Data Analytics at Logitech Snowflake + Tableau = #Winning

The Hadoop Ecosystem. EECS 4415 Big Data Systems. Tilemachos Pechlivanoglou

VOLTDB + HP VERTICA. page

Index. Scott Klein 2017 S. Klein, IoT Solutions in Microsoft s Azure IoT Suite, DOI /

Exam Questions

Architectural challenges for building a low latency, scalable multi-tenant data warehouse

Real4Test. Real IT Certification Exam Study materials/braindumps

Hadoop. Introduction / Overview

Leverage the Oracle Data Integration Platform Inside Azure and Amazon Cloud

Oracle Big Data SQL. Release 3.2. Rich SQL Processing on All Data

BI ENVIRONMENT PLANNING GUIDE

Data 101 Which DB, When Joe Yong Sr. Program Manager Microsoft Corp.

Big Data Infrastructure at Spotify

Overview. Prerequisites. Course Outline. Course Outline :: Apache Spark Development::

Talend Big Data Sandbox. Big Data Insights Cookbook

Data 101 Which DB, When. Joe Yong Azure SQL Data Warehouse, Program Management Microsoft Corp.

MAPR DATA GOVERNANCE WITHOUT COMPROMISE

Index A, B, C. Rank() function, steps, 199 Cloud services, 2 Comma-separated value (CSV), 27

1 Dulcian, Inc., 2001 All rights reserved. Oracle9i Data Warehouse Review. Agenda

Fast Innovation requires Fast IT

Data Science and Open Source Software. Iraklis Varlamis Assistant Professor Harokopio University of Athens

Accurate study guides, High passing rate! Testhorse provides update free of charge in one year!

Hadoop 2.x Core: YARN, Tez, and Spark. Hortonworks Inc All Rights Reserved

Ian Choy. Technology Solutions Professional

Technical Sheet NITRODB Time-Series Database

Case Study: Tata Communications Delivering a Truly Interactive Business Intelligence Experience on a Large Multi-Tenant Hadoop Cluster

How Apache Hadoop Complements Existing BI Systems. Dr. Amr Awadallah Founder, CTO Cloudera,

Azure Data Factory. Data Integration in the Cloud

Bring Context To Your Machine Data With Hadoop, RDBMS & Splunk

Implementing a Big Data Strategy PRASA Passenger Rail Agency of South Africa

Data Mining Concepts & Techniques

Combine Native SQL Flexibility with SAP HANA Platform Performance and Tools

Swimming in the Data Lake. Presented by Warner Chaves Moderated by Sander Stad

Azure Data Factory VS. SSIS. Reza Rad, Consultant, RADACAD

Partner Presentation Faster and Smarter Data Warehouses with Oracle OLAP 11g

SQL Server 2017 Power your entire data estate from on-premises to cloud

Data-Intensive Distributed Computing

Achieving Horizontal Scalability. Alain Houf Sales Engineer

Transcription:

Volume: 42 Questions Case Study: 1 Relecloud General Overview Relecloud is a social media company that processes hundreds of millions of social media posts per day and sells advertisements to several hundred companies. Relecloud has a Microsoft SQL Server database named DB1 that stores information about the advertisers. DB1 is hosted on a Microsoft Azure virtual machine. Physical locations Relecloud has two main offices. The offices we located in San Francisco and New York City. The offices connected to each other by using a site-to-site VPN. Each office connects directly to the Internet. Business model Relecloud modifies the pricing of its advertisements based on trending topics. Topics are considered to be trending if they generate many mentions in a specific country during a 15-minute time frame. The highest trending topics generate the highest advertising revenue. CTO statement Relecloud wants to deliver reports lo the advertisers by using Microsoft Power Bl. The reports will provide real-time data on trending topics, current advertising rates, and advertising costs for a given month. Relecloud will analyze the trending topics data, and then store the data in a new data warehouse for ad-hoc analysis. The data warehouse is expected to grow at a rate of 1 GB per hour or 8.7 terabytes (TB) per year. The data will be retained for five years for the purpose of long term trending. Requirements Business goals Management at Relecloud must be able to view which topics are trending to adjust advertising rates in near real-time. Planned changes Relecloud plans to implement a new streaming analytics platform that will report on trending topics. Relecloud plans to implement a data warehouse named DB2. General technical requirements Relecloud identifies the following technical requirements:. social media data must be analyzed to identify trending topics in real time. The use of Infrastructure as a Service (IaaS) platforms must minimized, whenever possible.

The real-time solution used to analyze the social media data must support selling up and down without service interruption. Technical requirements for advertisers Relecloud identifies the following technical requirements for the advertisers The advertisers must be able to see only their own data in the Power Bl reports. The advertisers must authenticate to Power Bl by using Azure Active Directory (Azure AD) credentials. The advertisers must be able to leverage existing Transact-SQL language knowledge when developing the real-time streaming solution. Members of the internal advertising sales team at Relecloud must be able to see only the sales data of the advertisers to which they are assigned. The Internal Relecloud advertising sales team must be prevented from inserting, updating, and deleting rows for the advertisers to which they are not assigned. The internal Relecloud advertising sales team must be able to use a text file to update the list of advertisers, and then to upload the file to Azure Blob storage. DB1 requirements Relecloud identifies the following requirements for DB1: Data generated by the streaming analytics platform must be stored in DB1. The user names of the advertisers must be mapped to CustomerID in a table named Table2. The advertisers in DB1 must be stored in a table named Table1 and must be refreshed nightly. The user names of the employees at Relecloud must be mapped to EmployeeID in a table named Table3. DB2 requirements Relecloud identifies the following requirements for DB2: DB2 must have minimal storage costs. DB2 must run load processes in parallel. DB2 must support massive parallel processing. DB2 must be able to store more than 40 TB of data. DB2 must support scaling up and down, as required. Data from DB1 must be archived in DB2 for long-term storage. All of the reports that are executed from DB2 must use aggregation. users must be able to pause DB2 when the data warehouse is not in use. users must be able to view previous versions of the data in DB2 by using aggregates. ETL requirements Relecloud identifies the following requirements for extract, transformation, and load (ETL): Data movement between DB1 and DB2 must occur each hour. An email alert must be generated when a failure of any type occurs during ETL processing. rls_table1

You execute the following code for a table named rls_table1. dbo.table1 You use the following code to create Table1. Streaming data The following is a sample of the Streaming data. Question: 1 Which technology should you recommend to meet the technical requirement for analyzing the social media data? A. Azure Stream Analytics B. Azure Data Lake Analytics C. Azure Machine Learning D. Azure HD lnsight Storm clusters A

Question: 2 DRAG DROP You need to implement a solution that meets the data refresh requirement tor DB1. Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order. Question: 3 DRAG DROP You need to create a query that identifies the trending topics. How should you complete the query? To answer, drag the appropriate values to the correct targets. Each value may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content

Question: 4 HOTSPOT You implement DB2. You need to configure the tables in DB2 to host the data from DB1. The solution must meet the requirements for DB2. Which type of table and history table storage should you use for the table? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one pint. Question: 5 DRAG DROP You need to implement rls_table1. Which code should you execute? To answer, drag the appropriate values to the correct targets. Each value may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.

Case Study: 2 Mix Questions Question: 1 Your company has multiple databases that contain millions of sales transactions. You plan to implement a data mining solution to identity purchasing fraud. You need to design a solution that mines 10 terabytes (TB) of sales data. The solution must meet the following requirements: Run the analysis to identify fraud once per week. Continue to receive new sales transactions while the analysis runs. Be able to stop computing services when the analysis is NOT running. Solution: You create a Cloudera Hadoop cluster on Microsoft Azure virtual machines. Does this meet the goal? A. Yes B. No A Question: 2 Your company has multiple databases that contain millions of sales transactions. You plan to implement a data mining solution to identity purchasing fraud. You need to design a solution that mines 10 terabytes (TB) of sales data. The solution must meet the following requirements: Run the analysis to identify fraud once per week. Continue to receive new sales transactions while the analysis runs. Be able to stop computing services when the analysis is NOT running. Solution: You create a Microsoft Azure Data Lake job. Does this meet the goal? A. Yes

B. No A Question: 3 Your company has multiple databases that contain millions of sales transactions. You plan to implement a data mining solution to identity purchasing fraud. You need to design a solution that mines 10 terabytes (TB) of sales data. The solution must meet the following requirements: Run the analysis to identify fraud once per week. Continue to receive new sales transactions while the analysis runs. Be able to stop computing services when the analysis is NOT running. Solution: You create a Microsoft Azure HD Insight cluster. Does this meet the goal? A. Yes B. No B Question: 4 HOTSPOT You are designing a solution based on the lambda architecture. The solution has the following layers; Batch Speed Serving You are planning the data ingestion process and the query execution. For each of the following statements, select Yes if the statement is true. Otherwise, select No.

Question: 5 DRAG DROP You have a web app that accepts user input, and then uses a Microsoft Azure Machine Learning model to predict a characteristic of the user. You need to perform the following operations: Track the number of web app users from month to month. Track the number of successful predictions made during the last minute. Create a dashboard showcasing the analytics tor the predictions and the web app usage. Which lambda layer should you query for each operation? To answer, drag the appropriate layers to the correct operations. Each layer may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content. Question: 6 HOTSPOT Your company has 2000 servers. You plan to aggregate all of the log files from the servers in a central repository that uses Microsoft Azure HD Insight. Each log file contains approximately one million records. All of the files use the.log file name extension. The following is a sample of the entries in the log files. 2017-02-03 20:26:41 SampleClass3 (ERROR) verbose detail for id 1527353937 In Apache Hive, you need to create a data definition and a query capturing tire number of records that have an error level of [ERROR]. What should you do? To answer, select the appropriate options in the answer area.

Question: 7 DRAG DROP You work for a telecommunications company that uses Microsoft Azure Stream Analytics. You have data related to incoming calls. You need to group the data in the following ways: Group A: Every five minutes for a duration of five minutes Group B: Every five minutes for a duration of 10 minutes Which type of window should you use for each group? To answer, drag the appropriate window types to the correct groups. Each window type may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.

Question: 8 HOTSPOT You have a Microsoft Azure Stream Analytics solution. You need to identify which types of windows must be used to group lite following types of events: Events that have random time intervals and are captured in a single fixed-size window Events that have random time intervals and are captured in overlapping windows Which window type should you identify for each event type? To answer, select the appropriate options in the answer area. Question: 9 You are designing an Apache HBase cluster on Microsoft Azure HD Insight. You need to identify which nodes are required for the cluster.