Which of the following statements are TRUE regarding the use of Data Click to load data into BigInsights? (Choose two.)
A. Big SQL cannot be used to access the data moved in by Data Click because the data is in Hive
B. You must import metadata for all sources and targets that you want to make available for Data Click activities
C. Connections from the relational database source to HDFS are discovered automatically from within Data Click
D. Hive tables are automatically created every time you run an activity that moves data from a relational database into HDFS
E. HBase tables are automatically created every time you ran an activity that moves data from a relational database into HDFS
Which of the following is TRUE about storing an Apache Spark object in serialized form?
A. It is advised to use Java serialization over Kryo serialization
B. Storing the object in serialized from will lead to faster access times
C. Storing the object in serialized from will lead to slower access times
D. All of the above
Which of the following statements regarding Big SQL is TRUE?
A. Big SQL doesn't support stored procedures
B. Big SQL can be deployed on a subset of data nodes in the BigInsights cluster
C. Big SQL provides a SQL-on-Hadoop environment based on map reduce
D. Only tables created or loaded via Big SQL can be accessed via Big SQL
Which of the following Hive data types is directly supported in Big SQL without any changes?
A. INT
B. STRING
C. STRUCT
D. BOOLEAN
Which of the following is not a capability of Pig?
A. Low-latency queries
B. Schemas are optional
C. Nested relational data model
D. A high level abstraction on top of MapReduce
Which one of the following is NOT provided by the SerDe interface?
A. SerDe interface has to be built using C or C++ language
B. Allows SQL-style queries across data that is often not appropriate for a relational database
C. Serializer takes a Java object that Big SQL has been working with, and turns it into a format that BigSQL can write to HDFS
D. Deserializer interface takes a string or binary representation of a record, and translates it into a Java
object that Big SQL can manipulate
What does the acronym "PCI" stand for in the phrase "PCI compliant"?
A. Payment Card Industry
B. Personal Credit and Income
C. Premium Credit Inspection
D. Proactive Controls Implementation
Which of the following statement is TRUE with BigSheets?
A. You can create any type of sheet from a parent workbook
B. You must create a child workbook in order to create a chart
C. You can delete a parent workbook without deleting the child workbooks
D. You must run the workbook on the data to get the full results of the analysis
When loading data into Big SQL, which statement is TRUE concerning the underlying storage mechanisms supported?
A. Big SQL supports .DB2 files
B. Big SQL supports Parquet files
C. Big SQL natively supports XML file format
D. Big SQL supports file stored in PDF format
Which of the following is not a data-processing operations that is supported in Pig Latin?
A. filter
B. joins
C. group by
D. logistic regression
What are the key elements that IBM Big Match Probabilistic Matching Engine leverages?
A. Wildcard and sorting
B. Wildcard and phonetics
C. Phonetics and searching
D. Phonetics and nicknames
Which of the following must happen before the Big SQL EXPLAIN command can execute?
A. Run the ANALYZE command
B. Set the COMPATIBILITY_MODE global variable
C. Execute the SET HADOOP PROPERTY command
D. Call the SYSPROC.SYSINSTALLOBJECTS procedure
What should you use to run Big SQL queries that require dynamic variables or scheduling?
A. JSqsh
B. Eclipse
C. BigInsights Console
D. BigInsights Applications
A large retailer is planning to create a data cloud using BigInsights. They would like to run queries using data that is stored in BigInsights as well as other databases like DB2 and Oracle.
Which one of the following would enable them to write the queries without having to move data into one common data store?
A. Hive
B. HBase
C. Big SQL
D. BigSheets
How many Job Trackers can be found in a MapReduce v1 cluster?
A. One per cluster
B. One per data node
C. One for each Mapper
D. One for each Reducer