Which methods of Kafka deployment on Oracle Cloud does NOT provide SSH access?
A. Oracle Event Hub Cloud Service Dedicated associated with Big Data Cloud
B. Oracle Event Hub Cloud Service (Multi Tenant Version)
C. Oracle Event Hub Cloud Service-Dedicated (Kafka as a service)
D. Customer installed Kafka on Oracle Cloud Infrastructure
You have easily and successfully created clusters with the Oracle Big Data Cloud wizard. You want to
create a cluster that will be very specific to the needs of your business.
How would you customize Oracle Big Data Cloud clusters during provisioning?
A. by using Stack Manager
B. by using Oracle Enterprise Manager
C. by using Platform Service Manager UI
D. by using a Bootstrap script
You are in a project with teams from other departments of the company. It is necessary to collaborate not only creative ideas but also the data with the other departments. In the Big Data Cloud Service, you need to copy data that is in very large files between HDFS on your cluster to their cloud storage.
Which utility is more efficient for copying large data files?
A. odcp
B. ftp
C. fastcopy
D. scp
What is the difference between permanent nodes and edge nodes?
A. Permanent nodes cannot be stopped, whereas you can start and stop edge nodes.
B. Permanent nodes are for the life of the cluster, whereas edge nodes are temporary for the duration of processing the data Permanent nodes contain your Hadoop data, but edge nodes do not have Hadoop data.
C. Permanent nodes contain your Hadoop data, but edge nodes do not have Hadoop data.
D. Permanent nodes contain your Hadoop data, but edge nodes give you the "edge" in processing your data with more processors.
What is the result of the FLATMAP () function in Spark?
A. It always returns a new RDD by passing the supplied function used to filter the results.
B. It always returns a new RDD that contains elements in the source dataset and the argument.
C. It always returns an RDD with 0, 1, or more elements.
D. It always returns an RDD with identical size of the input RDD.
How can you collect and move Syslog data with HDFS without any data loss?
A. Use the HDFS copy command as long as you have necessary HDFS and Cluster Access Privileges.
B. Use FLUME with Memory channel and Sink.
C. Use FLUME with In File channel
D. Use PIG with error checking enabled
ABC Company maintains two Hadoop clusters, one for test and one for production. One of the developers wants to copy a data file from the test to the production environment.
Which two are possible options to accomplish this?
A. Use Flume to ingest the file again and deliver it to production.
B. Back up the file to tape and restore it to the production cluster.
C. Use an FTP command to copy files between the test and production environments.
D. Use the hdfs dfs -cp command to copy the file from test to production.
Oracle's unified data management platform provides analysts with the ability to execute a single SQL statement across many data stores.
Which three are data stores the SQL can be executed in?
A. NoSQL databases
B. Hadoop clusters
C. Object-oriented databases
D. Directory service store
E. Relational databases
Which three statements are true about Whitelists?
A. It is a list of all approved, enabled clusters.
B. It is used to filter network traffic to services in the cluster.
C. It specifies if clients for specific services on specific ports are denied access.
D. It is list of clusters accessible to you.
E. It is an index of approved entities.
Given: A Kerberos principal is a unique identity to which Kerberos can assign tickets, and a realm is where the Kerberos database is stored.
What is the realm used by Big Data Cloud Service?
A. bigdatacloudservice.oracle.com
B. bdcs.oracle.com
C. bdacoudservice.oracle.com
D. bigdata.oracle.com
Which can you add more of to scale an Event Hub Cloud Service Topic?
A. Zookeeper nodes
B. Kafka brokers
C. Partitions
D. Big data nodes
Which two are the supported SDKs to manage data in Oracle Big Data Manager?
A. Scala
B. HTML
C. Ruby
D. Python
E. Java
What does Schema On Read represent?
A. Ingest all data and create structure per query requirements
B. Create table structure, ingest data and query
C. In-memory algorithm that defines the schema on the data
D. Specialized schema optimized for read
Which three statements are true about YARN?
A. It handles and schedules resource requests.
B. Yarn can run applications that follow MapReduce and Non MapReduce Model like Graph Processing.
C. Yarn does NOT support running multiple applications in a single Hadoop Cluster.
D. It provides Daemons and API to develop generic distributed applications.
E. Yarn model Functionality is identical to Apache Hadoop MapReduce (MR1).
F. Yarn combines the two major functionalities of Job tracker: Resource Management and Job Scheduling
You have data stored in the Hadoop cluster. You wish to use "R" to process the data in Hadoop and interface it with your Oracle database with statistical algorithms.
Which function should you use to perform a linear regression model on the data?
A. orch.lm
B. orch.lmf
C. orch.lr
D. orch.nmf