Question: 1 You need to place the results of a PigLatin script into an HDFS output directory. What is the correct syntax in Apache Pig?

Similar documents
Exam Questions 1z0-449

Actual4Test. Actual4test - actual test exam dumps-pass for IT exams

IT Certification Exams Provider! Weofferfreeupdateserviceforoneyear! h ps://

1z0-449.exam. Number: 1z0-449 Passing Score: 800 Time Limit: 120 min File Version: Oracle. 1z0-449

1Z Oracle Big Data 2017 Implementation Essentials Exam Summary Syllabus Questions

Oracle. Oracle Big Data 2017 Implementation Essentials. 1z Version: Demo. [ Total Questions: 10] Web:

Oracle 1Z Oracle Big Data 2017 Implementation Essentials.

Oracle Big Data Fundamentals Ed 1

Oracle Big Data Fundamentals Ed 2

Oracle Big Data Connectors

Oracle NoSQL Database Enterprise Edition, Version 18.1

We are ready to serve Latest Testing Trends, Are you ready to learn?? New Batches Info

Hadoop. Course Duration: 25 days (60 hours duration). Bigdata Fundamentals. Day1: (2hours)

Oracle Big Data Appliance

Oracle Big Data. A NA LYT ICS A ND MA NAG E MENT.

Big Data Hadoop Stack

Big Data Hadoop Developer Course Content. Big Data Hadoop Developer - The Complete Course Course Duration: 45 Hours

Securing the Oracle BDA - 1

Big Data Architect.

Introduction to BigData, Hadoop:-

We are ready to serve Latest Testing Trends, Are you ready to learn.?? New Batches Info

Oracle Big Data Appliance

Innovatus Technologies

Security and Performance advances with Oracle Big Data SQL

Building an Integrated Big Data & Analytics Infrastructure September 25, 2012 Robert Stackowiak, Vice President Data Systems Architecture Oracle

Big Data Analytics using Apache Hadoop and Spark with Scala

Overview. : Cloudera Data Analyst Training. Course Outline :: Cloudera Data Analyst Training::

Hadoop Development Introduction

MapR Enterprise Hadoop

Oracle Big Data Appliance

Blended Learning Outline: Cloudera Data Analyst Training (171219a)

CIS 612 Advanced Topics in Database Big Data Project Lawrence Ni, Priya Patil, James Tench

CISC 7610 Lecture 2b The beginnings of NoSQL

Oracle NoSQL Database Enterprise Edition, Version 18.1

The Hadoop Ecosystem. EECS 4415 Big Data Systems. Tilemachos Pechlivanoglou

Narration Script for ODI Adapter for Hadoop estudy

International Journal of Advance Engineering and Research Development. A Study: Hadoop Framework

Integrating Big Data with Oracle Data Integrator 12c ( )

Bring Context To Your Machine Data With Hadoop, RDBMS & Splunk

Hortonworks Certified Developer (HDPCD Exam) Training Program

Certified Big Data and Hadoop Course Curriculum

Lecture 7 (03/12, 03/14): Hive and Impala Decisions, Operations & Information Technologies Robert H. Smith School of Business Spring, 2018

How Apache Hadoop Complements Existing BI Systems. Dr. Amr Awadallah Founder, CTO Cloudera,

Certified Big Data Hadoop and Spark Scala Course Curriculum

Hadoop An Overview. - Socrates CCDH

Hadoop. Introduction / Overview

HADOOP FRAMEWORK FOR BIG DATA

Copyright 2013, Oracle and/or its affiliates. All rights reserved.

CERTIFICATE IN SOFTWARE DEVELOPMENT LIFE CYCLE IN BIG DATA AND BUSINESS INTELLIGENCE (SDLC-BD & BI)

This is a brief tutorial that explains how to make use of Sqoop in Hadoop ecosystem.

50 Must Read Hadoop Interview Questions & Answers

<Insert Picture Here> Introduction to Big Data Technology

SOLUTION TRACK Finding the Needle in a Big Data Innovator & Problem Solver Cloudera

Hadoop 2.x Core: YARN, Tez, and Spark. Hortonworks Inc All Rights Reserved

Things Every Oracle DBA Needs to Know about the Hadoop Ecosystem. Zohar Elkayam

Big Data Syllabus. Understanding big data and Hadoop. Limitations and Solutions of existing Data Analytics Architecture

Microsoft Big Data and Hadoop

Hadoop. copyright 2011 Trainologic LTD

Oracle Big Data Connectors

Copyright 2012, Oracle and/or its affiliates. All rights reserved.

Introduction to Hadoop. High Availability Scaling Advantages and Challenges. Introduction to Big Data

Hadoop-PR Hortonworks Certified Apache Hadoop 2.0 Developer (Pig and Hive Developer)

Hadoop. Introduction to BIGDATA and HADOOP

Hadoop and HDFS Overview. Madhu Ankam

HADOOP COURSE CONTENT (HADOOP-1.X, 2.X & 3.X) (Development, Administration & REAL TIME Projects Implementation)

Big Data with Hadoop Ecosystem

Hadoop is supplemented by an ecosystem of open source projects IBM Corporation. How to Analyze Large Data Sets in Hadoop

Importing and Exporting Data Between Hadoop and MySQL

Topics. Big Data Analytics What is and Why Hadoop? Comparison to other technologies Hadoop architecture Hadoop ecosystem Hadoop usage examples

Big Data Development HADOOP Training - Workshop. FEB 12 to (5 days) 9 am to 5 pm HOTEL DUBAI GRAND DUBAI

CONSOLIDATING RISK MANAGEMENT AND REGULATORY COMPLIANCE APPLICATIONS USING A UNIFIED DATA PLATFORM

Oracle Big Data Appliance

Hadoop & Big Data Analytics Complete Practical & Real-time Training

April Final Quiz COSC MapReduce Programming a) Explain briefly the main ideas and components of the MapReduce programming model.

Part 1 Configuring Oracle Big Data SQL

Oracle NoSQL Database and Cisco- Collaboration that produces results. 1 Copyright 2011, Oracle and/or its affiliates. All rights reserved.

Big Data Hadoop Course Content

Safe Harbor Statement

Oracle Big Data Appliance

Big Data com Hadoop. VIII Sessão - SQL Bahia. Impala, Hive e Spark. Diógenes Pires 03/03/2018

MODERN BIG DATA DESIGN PATTERNS CASE DRIVEN DESINGS

Big Data. Big Data Analyst. Big Data Engineer. Big Data Architect

Oracle Big Data Connectors

docs.hortonworks.com

User's Guide Release 4 (4.1)

Big Data Technology Ecosystem. Mark Burnette Pentaho Director Sales Engineering, Hitachi Vantara

Delving Deep into Hadoop Course Contents Introduction to Hadoop and Architecture

International Journal of Advance Engineering and Research Development. A study based on Cloudera's distribution of Hadoop technologies for big data"

exam. Microsoft Perform Data Engineering on Microsoft Azure HDInsight. Version 1.0

Oracle Data Integrator 12c: Integration and Administration

Stages of Data Processing

Oracle. NoSQL Database Concepts Manual. 12c Release 1

Oracle BDA: Working With Mammoth - 1

基于 Hadoop 和 RDBMS 的 Oracle 大数据分析 Corey Wei 技术顾问甲骨文公司. 版权所有 2014,Oracle 和 / 或其关联公司 保留所有权利

Impala. A Modern, Open Source SQL Engine for Hadoop. Yogesh Chockalingam

Percona Live September 21-23, 2015 Mövenpick Hotel Amsterdam

Introduction to HDFS and MapReduce

Hadoop Online Training

Chase Wu New Jersey Institute of Technology

Quick Deployment Step-by-step instructions to deploy Oracle Big Data Lite Virtual Machine

Transcription:

Volume: 72 Questions Question: 1 You need to place the results of a PigLatin script into an HDFS output directory. What is the correct syntax in Apache Pig? A. update hdfs set D as./output ; B. store D into./output ; C. place D into./output ; D. write D as./output ; E. hdfsstore D into./output ; Question: 2 The Hadoop NameNode is running on port #3001, the DataNode on port #4001, the KVStore agent on port #5001, and the replication node on port #6001. All the services are running on localhost. What is the valid syntax to create an external table in Hive and query data from the NoSQL Database? A. CREATE EXTERNAL TABLE IF NOT EXISTSMOVIE( id INT,original_tit1e STRING,overview STRING)STORED BY 'oracle.kv.hadoop.hive.table.tablestoragehandler'tblproperties ("oracle.kv.kvstore"="kvscore","oracle.kv.hosts"="localhost:3001","oracle.kv.hadoop.hosts" ="localhost","oracle.kv.tablename"= MOVIE"); B. CREATE EXTERNAL TABLE IF NOT EXISTSMOVIE( id INT,original_title STRING,overview STRING)STORED BY 'oracle.kv.hadoop.hive.table.tablestoragehandler'tblproperties ("oracle.kv.kvstore "=" kvstore","oracle.kv.hosts"="localhost:5001","oracle.kv.hadoop.hosts" ="localhost","oracle.kv.tab1ename"="movie"); C. CREATE EXTERNAL TABLE IF NOT EXISTSMOVIE( id INT,original_title STRING,overview STRING)STORED BY 'oracle,kv.hadoop.hive.table.tablestoragehandler'tblproperties ("oracle.kv.kvstore"="kvstore","oracle.kv.hosts"="localhost:4001","oracle.kv.hadoop.hosts"

= "localhost","oracle.kv.tab1ename"="movie"); D. CREATE EXTERNAL TABLE IF NOT EXISTSMOVIE( id INT,original_title STRING,overview STRING)STORED BY 'oracle,kv.hadoop.hive.table.tablestoragehandler'tblproperties ("oracle.kv.kvstore"="kvstore","oracle.kv.hosts"="localhost:6001","oracle.kv.hadoop.hosts"= "localhost","oracle.kv.tab1ename"="movie"); Answer: C Question: 3 You need to create an architecture for your customer s Oracle NoSQL KVStore. The customer needs to store clinical and non-clinical data together but only the clinical data is mission critical. How can both types of data exist in the same KVStore? A. Store the clinical data on the master node and the non-clinical data on the replica nodes. B. Store the two types of data in separate partitions on highly available storage. C. Store the two types of data in two separate KVStore units and create database aliases to mimic one KVStore. D. Store the two types of data with differing consistency policies. Question: 4 Your customer is spending a lot of money on archiving data to comply with government regulations to retain data for 10 years. How should you reduce your customer s archival costs? A. Denormalize the data. B. Offload the data into Hadoop. C. Use Oracle Data Integrator to improve performance. D. Move the data into the warehousing database.

Question: 5 Your customer keeps getting an error when writing a key/value pair to a NoSQL replica. What is causing the error? A. The master may be in read-only mode and as result, writes to replicas are not being allowed. B. The replica may be out of sync with the master and is not able to maintain consistency. C. The writes must be done to the master. D. The replica is in read-only mode. E. The data file for the replica is corrupt. Answer: C Question: 6 What two actions do the following commands perform in the Oracle R Advanced Analytics for Hadoop Connector? (Choose two.) ore.connect (type= HIVE ) ore.attach () A. Connect to Hive. B. Attach the Hadoop libraries to R. C. Attach the current environment to the search path of R. D. Connect to NoSQL via Hive. Answer: A,C Question: 7 Your customer needs to access Hive tables, HDFS, and Data Pump format files as the data source. When installing and configuring the Oracle SQL Connector for HDFS, which two locations must be installed or configured to fulfill this requirement? (Choose two.) A. Hadoop cluster B. HiveQL

C. Impala D. Oracle database E. Oracle NoSQL Answer: A,D Question: 8 Which statement is true about the NameNode in Hadoop? A. A query in Hadoop requires a MapReduce job to be run so the NameNode gets the location of the data from the JobTracker. B. If the NameNode goes down and a secondary NameNode has not been defined, the cluster is still accessible. C. When loading data, the NameNode tells the client or program where to write the data. D. All data passes through the NameNode; so if it is not sized properly, it could be a potential bottleneck. Question: 9 What does the following line do in Apache Pig? products = LOAD /user/oracle/products AS (prod_id, item); A. The products table is loaded by using data pump with prod_id and item. B. The LOAD table is populated with prod_id and item. C. The contents of /user/oracle/products are loaded as tuples and aliased to products. D. The contents of /user/oracle/products are dumped to the screen. Answer: C Question: 10 Your customer needs to analyze large numbers of log files after combining them with customer

profile and interaction data. The customer wants to be able to analyze these files back to their go-live date, which was last October. Which solution should you choose to accomplish this? A. Oracle Database B. Hadoop C. Berkeley DB D. Oracle NoSQL Question: 11 Your customer wants to implement the Oracle Table Access for a Hadoop Connector (OTA4H) to connect the Hadooop cluster to the Oracle database. The customer s Oracle Database is a Real Application Clusters database. How should you configure the database listener? A. The database listener is configured to route requests to the designated "hadoop" node in the listener s service definition. B. The database listener is configured to route the Hadoop connection requests to the available nodes randomly, provided that the RAC-aware service name is specified in the JDBC URL. C. The database listener is configured to route the Hadoop connection requests to the first node the RAC cluster because clustered environments are not inherently supported. D. Connecting to a RAC database is not supported. Question: 12 Your customer completed all the Kerberos installation prerequisites when the Big Data Appliance was set up. However, when the customer tries to use Kerberos authentication, it gets an error. Which command did the customer fail to run? A. install.sh option kerberos

B. emcli enable kerberos C. bdacli enable kerberos D. bdasetup kerberos Answer: C Question: 13 Your customer has had a security breach on its running Big Data Appliance and needs to quickly enable security for Hive and Impala. What is the quickest way to enable the Sentry service? A. Execute bdacli sentry enable. B. Use the Mammoth utility during installation. C. Sentry is enabled by default on the Big Data Appliance. D. Use the Cloudera Manager. Answer: D Question: 14 Which user in the following set of ACL entries has read, write and execute permissions? user::rwx group::r-x other::r-x default:user::rwx default:user:bruce:rwx #effective:r-x default:group::r-x default:group:sales:rwx #effective:r-x default:mask::r-x default:other::r-x A. sales B. bruce C. all users

D. hdfs E. user Answer: E Question: 15 Identify two valid steps for setting up a Hive-to-Hive transformation by using Oracle Data Integrator. (Choose two.) A. Ensure that the Hive server is up and running. B. Create a Logical Schema object. C. Configure ODI by using the mammoth utility. D. Ensure that Apache Sentry is configured. Answer: A,B Question: 16 Your customer wants you to set up ODI by using the IKM SQL to Hive module and overwrite existing Hive tables with the customer s new clinical data. What parameter must be set to true? A. SQOOP_ODI_OVERWRITE B. OVERWRITE_HIVE_TABLE C. OVERWRITE_HDFS_TABLE D. CREATE_HIVE_TABLE Answer: D Question: 17 How does increasing the number of storage nodes and shards impact the efficiency of Oracle NoSQL Database? A. The number of shards or storage nodes does not impact performance.

B. Having more shards reduces the write throughput because of increased node forwarding. C. Having more shards results in reduced read throughput because of increased node forwarding. D. Having more shards increases the write throughput because more master nodes are available for writes. Answer: D Question: 18 How does the Sqoop Server handle a job that was discontinued by a restart? A. by sending an alert and waiting to resume the job B. by continuing the loading of data C. by stopping the loading of data D. by rolling back the load Question: 19 For Oracle R Advanced Analytics for Hadoop to access the data stored in HDFS, what must the input files comply with? A. All the input files for a MapReduce job must be stored in one directory as part of one logical file. B. Only input files with an underscore (_) are processed. C. Only files with tab delimiters are supported. D. Oracle R Advanced Analytics does not have any restrictions on input files. Answer: A Question: 20 What is the main purpose of the Oracle Loader for Hadoop (OLH) Connector?

A. runs transformations expressed in XQuery by translating them into a series of MapReduce jobs that are executed in parallel on a Hadoop cluster B. pre-partitions, sorts, and transforms data into an Oracle ready format on Hadoop and loads it into the Oracle database C. accesses and analyzes data in place on HDFS by using external tables D. performs scalable joins between Hadoop and Oracle Database data E. provides a SQL-like interface to data that is stored in HDFS F. is the single SQL point-of-entry to access all data Question: 21 What kind of workload is MapReduce designed to handle? A. batch processing B. interactive C. computational D. real time E. commodity Answer: A Question: 22 Your customer has three XML files in HDFS with the following contents. Each XML file contains comments made by users on a specific day. Each comment can have zero or more likes from other users. The customer wants you to query this data and load it into the Oracle Database on Exadata. How should you parse this data?

A. by creating a table in Hive and using MapReduce to parse the XML data by column B. by configuring the Oracle SQL Connector for HDFS and parsing by using SerDe C. by using the XML file module in the Oracle XQuery for Hadoop Connector D. by using the built-in functions for reading JSON in the Oracle XQuery for Hadoop Connector Question: 23 Your customer needs the data that is generated from social media such as Facebook and Twitter, and the customer s website to be consumed and sent to an HDFS directory for analysis by the marketing team. Identify the architecture that you should configure. A. multiple flume agents with collectors that output to a logger that writes to the Oracle Loader for Hadoop agent B. multiple flume agents with sinks that write to a consolidated source with a sink to the customer's HDFS directory C. a single flume agent that collects data from the customer's website, which is connected to both Facebook and Twitter, and writes via the collector to the customer's HDFS directory