Volume: 72 Questions Question: 1 You need to place the results of a PigLatin script into an HDFS output directory. What is the correct syntax in Apache Pig? A. update hdfs set D as./output ; B. store D into./output ; C. place D into./output ; D. write D as./output ; E. hdfsstore D into./output ; Question: 2 The Hadoop NameNode is running on port #3001, the DataNode on port #4001, the KVStore agent on port #5001, and the replication node on port #6001. All the services are running on localhost. What is the valid syntax to create an external table in Hive and query data from the NoSQL Database? A. CREATE EXTERNAL TABLE IF NOT EXISTSMOVIE( id INT,original_tit1e STRING,overview STRING)STORED BY 'oracle.kv.hadoop.hive.table.tablestoragehandler'tblproperties ("oracle.kv.kvstore"="kvscore","oracle.kv.hosts"="localhost:3001","oracle.kv.hadoop.hosts" ="localhost","oracle.kv.tablename"= MOVIE"); B. CREATE EXTERNAL TABLE IF NOT EXISTSMOVIE( id INT,original_title STRING,overview STRING)STORED BY 'oracle.kv.hadoop.hive.table.tablestoragehandler'tblproperties ("oracle.kv.kvstore "=" kvstore","oracle.kv.hosts"="localhost:5001","oracle.kv.hadoop.hosts" ="localhost","oracle.kv.tab1ename"="movie"); C. CREATE EXTERNAL TABLE IF NOT EXISTSMOVIE( id INT,original_title STRING,overview STRING)STORED BY 'oracle,kv.hadoop.hive.table.tablestoragehandler'tblproperties ("oracle.kv.kvstore"="kvstore","oracle.kv.hosts"="localhost:4001","oracle.kv.hadoop.hosts"
= "localhost","oracle.kv.tab1ename"="movie"); D. CREATE EXTERNAL TABLE IF NOT EXISTSMOVIE( id INT,original_title STRING,overview STRING)STORED BY 'oracle,kv.hadoop.hive.table.tablestoragehandler'tblproperties ("oracle.kv.kvstore"="kvstore","oracle.kv.hosts"="localhost:6001","oracle.kv.hadoop.hosts"= "localhost","oracle.kv.tab1ename"="movie"); Answer: C Question: 3 You need to create an architecture for your customer s Oracle NoSQL KVStore. The customer needs to store clinical and non-clinical data together but only the clinical data is mission critical. How can both types of data exist in the same KVStore? A. Store the clinical data on the master node and the non-clinical data on the replica nodes. B. Store the two types of data in separate partitions on highly available storage. C. Store the two types of data in two separate KVStore units and create database aliases to mimic one KVStore. D. Store the two types of data with differing consistency policies. Question: 4 Your customer is spending a lot of money on archiving data to comply with government regulations to retain data for 10 years. How should you reduce your customer s archival costs? A. Denormalize the data. B. Offload the data into Hadoop. C. Use Oracle Data Integrator to improve performance. D. Move the data into the warehousing database.
Question: 5 Your customer keeps getting an error when writing a key/value pair to a NoSQL replica. What is causing the error? A. The master may be in read-only mode and as result, writes to replicas are not being allowed. B. The replica may be out of sync with the master and is not able to maintain consistency. C. The writes must be done to the master. D. The replica is in read-only mode. E. The data file for the replica is corrupt. Answer: C Question: 6 What two actions do the following commands perform in the Oracle R Advanced Analytics for Hadoop Connector? (Choose two.) ore.connect (type= HIVE ) ore.attach () A. Connect to Hive. B. Attach the Hadoop libraries to R. C. Attach the current environment to the search path of R. D. Connect to NoSQL via Hive. Answer: A,C Question: 7 Your customer needs to access Hive tables, HDFS, and Data Pump format files as the data source. When installing and configuring the Oracle SQL Connector for HDFS, which two locations must be installed or configured to fulfill this requirement? (Choose two.) A. Hadoop cluster B. HiveQL
C. Impala D. Oracle database E. Oracle NoSQL Answer: A,D Question: 8 Which statement is true about the NameNode in Hadoop? A. A query in Hadoop requires a MapReduce job to be run so the NameNode gets the location of the data from the JobTracker. B. If the NameNode goes down and a secondary NameNode has not been defined, the cluster is still accessible. C. When loading data, the NameNode tells the client or program where to write the data. D. All data passes through the NameNode; so if it is not sized properly, it could be a potential bottleneck. Question: 9 What does the following line do in Apache Pig? products = LOAD /user/oracle/products AS (prod_id, item); A. The products table is loaded by using data pump with prod_id and item. B. The LOAD table is populated with prod_id and item. C. The contents of /user/oracle/products are loaded as tuples and aliased to products. D. The contents of /user/oracle/products are dumped to the screen. Answer: C Question: 10 Your customer needs to analyze large numbers of log files after combining them with customer
profile and interaction data. The customer wants to be able to analyze these files back to their go-live date, which was last October. Which solution should you choose to accomplish this? A. Oracle Database B. Hadoop C. Berkeley DB D. Oracle NoSQL Question: 11 Your customer wants to implement the Oracle Table Access for a Hadoop Connector (OTA4H) to connect the Hadooop cluster to the Oracle database. The customer s Oracle Database is a Real Application Clusters database. How should you configure the database listener? A. The database listener is configured to route requests to the designated "hadoop" node in the listener s service definition. B. The database listener is configured to route the Hadoop connection requests to the available nodes randomly, provided that the RAC-aware service name is specified in the JDBC URL. C. The database listener is configured to route the Hadoop connection requests to the first node the RAC cluster because clustered environments are not inherently supported. D. Connecting to a RAC database is not supported. Question: 12 Your customer completed all the Kerberos installation prerequisites when the Big Data Appliance was set up. However, when the customer tries to use Kerberos authentication, it gets an error. Which command did the customer fail to run? A. install.sh option kerberos
B. emcli enable kerberos C. bdacli enable kerberos D. bdasetup kerberos Answer: C Question: 13 Your customer has had a security breach on its running Big Data Appliance and needs to quickly enable security for Hive and Impala. What is the quickest way to enable the Sentry service? A. Execute bdacli sentry enable. B. Use the Mammoth utility during installation. C. Sentry is enabled by default on the Big Data Appliance. D. Use the Cloudera Manager. Answer: D Question: 14 Which user in the following set of ACL entries has read, write and execute permissions? user::rwx group::r-x other::r-x default:user::rwx default:user:bruce:rwx #effective:r-x default:group::r-x default:group:sales:rwx #effective:r-x default:mask::r-x default:other::r-x A. sales B. bruce C. all users
D. hdfs E. user Answer: E Question: 15 Identify two valid steps for setting up a Hive-to-Hive transformation by using Oracle Data Integrator. (Choose two.) A. Ensure that the Hive server is up and running. B. Create a Logical Schema object. C. Configure ODI by using the mammoth utility. D. Ensure that Apache Sentry is configured. Answer: A,B Question: 16 Your customer wants you to set up ODI by using the IKM SQL to Hive module and overwrite existing Hive tables with the customer s new clinical data. What parameter must be set to true? A. SQOOP_ODI_OVERWRITE B. OVERWRITE_HIVE_TABLE C. OVERWRITE_HDFS_TABLE D. CREATE_HIVE_TABLE Answer: D Question: 17 How does increasing the number of storage nodes and shards impact the efficiency of Oracle NoSQL Database? A. The number of shards or storage nodes does not impact performance.
B. Having more shards reduces the write throughput because of increased node forwarding. C. Having more shards results in reduced read throughput because of increased node forwarding. D. Having more shards increases the write throughput because more master nodes are available for writes. Answer: D Question: 18 How does the Sqoop Server handle a job that was discontinued by a restart? A. by sending an alert and waiting to resume the job B. by continuing the loading of data C. by stopping the loading of data D. by rolling back the load Question: 19 For Oracle R Advanced Analytics for Hadoop to access the data stored in HDFS, what must the input files comply with? A. All the input files for a MapReduce job must be stored in one directory as part of one logical file. B. Only input files with an underscore (_) are processed. C. Only files with tab delimiters are supported. D. Oracle R Advanced Analytics does not have any restrictions on input files. Answer: A Question: 20 What is the main purpose of the Oracle Loader for Hadoop (OLH) Connector?
A. runs transformations expressed in XQuery by translating them into a series of MapReduce jobs that are executed in parallel on a Hadoop cluster B. pre-partitions, sorts, and transforms data into an Oracle ready format on Hadoop and loads it into the Oracle database C. accesses and analyzes data in place on HDFS by using external tables D. performs scalable joins between Hadoop and Oracle Database data E. provides a SQL-like interface to data that is stored in HDFS F. is the single SQL point-of-entry to access all data Question: 21 What kind of workload is MapReduce designed to handle? A. batch processing B. interactive C. computational D. real time E. commodity Answer: A Question: 22 Your customer has three XML files in HDFS with the following contents. Each XML file contains comments made by users on a specific day. Each comment can have zero or more likes from other users. The customer wants you to query this data and load it into the Oracle Database on Exadata. How should you parse this data?
A. by creating a table in Hive and using MapReduce to parse the XML data by column B. by configuring the Oracle SQL Connector for HDFS and parsing by using SerDe C. by using the XML file module in the Oracle XQuery for Hadoop Connector D. by using the built-in functions for reading JSON in the Oracle XQuery for Hadoop Connector Question: 23 Your customer needs the data that is generated from social media such as Facebook and Twitter, and the customer s website to be consumed and sent to an HDFS directory for analysis by the marketing team. Identify the architecture that you should configure. A. multiple flume agents with collectors that output to a logger that writes to the Oracle Loader for Hadoop agent B. multiple flume agents with sinks that write to a consolidated source with a sink to the customer's HDFS directory C. a single flume agent that collects data from the customer's website, which is connected to both Facebook and Twitter, and writes via the collector to the customer's HDFS directory