You are in a project with teams from other departments of the company. It is necessary to collaborate not only creative ideas but also the data with the other departments. In the Big Data Cloud Service, you need to copy data that is in very large files between HDFS on your cluster to their cloud storage.
Which utility is more efficient for copying large data files?
A. odcp
B. ftp
C. fastcopy
D. scp
Oracle Data Integrator for Big Data offers customers with Enterprise big data Integration.
What component does Oracle Data Integrator for Big Data use to give you the ability to solve your most
complex and time-sensitive data transformation and data movement challenges?
A. RDD
B. Knowledge modules
C. Predefined MapReduce job for data transformation
D. Package scripts
What is the result of the FLATMAP () function in Spark?
A. It always returns a new RDD by passing the supplied function used to filter the results.
B. It always returns a new RDD that contains elements in the source dataset and the argument.
C. It always returns an RDD with 0, 1, or more elements.
D. It always returns an RDD with identical size of the input RDD.
As the Big Data Cloud Service Administrator, you need to access one of the cluster nodes. Using secure shell, what are the two things you will need in order to access the node?
A. Name of the cluster node
B. Private ssh key pair
C. Name of the service instance
D. IP Address of the cluster node
E. Public ssh key pair
Which three statements are true about Whitelists?
A. It is a list of all approved, enabled clusters.
B. It is used to filter network traffic to services in the cluster.
C. It specifies if clients for specific services on specific ports are denied access.
D. It is list of clusters accessible to you.
E. It is an index of approved entities.
Which two are possible integration options that can be used to integrate data between Oracle database table with LOB columns and data stored in Hadoop HDFS?
A. Use SQOOP JDBC based imports
B. Use FLUME with database source along with HDFS file Source agents to integrate data
C. Use SQOOP with -direct option
D. Use Oracle SQL connector for Hadoop
A financial company uses Hadoop cluster to store, process, and analyze all the stock trading data. They use a map reduce job to generate daily reports for their clients by 6 a.m. every day using the data on Hadoop cluster. One of the data nodes in the cluster crashed due to hard disk failure.
What will be the impact?
A. Restore the data from backup to the new node and rerun the report
B. HDFS replication factor ensures that report generation process runs to completion
C. Hadoop can reboot the failed node automatically and rectify the faulty disk
D. Downtime will be needed to replace the failed hard disk and rerun the report
On which open-source processing engine is Oracle Big Data Cloud (BDC) Strategy based on?
A. Spark
B. Apache Flink Scalable Batch and Stream Data Processing
C. Apache Apex
D. Apache Storm
What does Schema On Read represent?
A. Ingest all data and create structure per query requirements
B. Create table structure, ingest data and query
C. In-memory algorithm that defines the schema on the data
D. Specialized schema optimized for read
You have data originating from different sources and the data is not structured the same way. You need to combine the data from those heterogeneous sources into integrated, consistent, and unambiguous information so you can discover the relationships of the data.
Which Oracle Big Data Connector should you use?
A. Oracle R Advanced Analytics for Hadoop
B. Oracle Big Data Spatial and Graph
C. Oracle Non-relational Data Integrator
D. Oracle SQL Connector for Hadoop Distributed File System