▸ Hadoop | LinkedIn Skill Assessment Quiz Solutions
Recommended Big Data Hadoop Courses:
Partitioner controls the partitioning of what data?
- final keys
- final values
- intermediate keys
- intermediate values
SQL Windowing functions are implemented in Hive using which keywords?
- UNION DISTINCT, RANK
- OVER, RANK
- OVER, EXCEPT
- UNION DISTINCT, RANK
Rather than adding a Secondary Sort to a slow Reduce job, it is Hadoop best practice to perform which optimization?
- Add a partitioned shuffle to the Map job.
- Add a partitioned shuffle to the Reduce job.
- Break the Reduce job into multiple, chained Reduce jobs.
- Break the Reduce job into multiple, chained Map jobs.
Hadoop Auth enforces authentication on protected resources. Once authentication has been established, it sets what type of authenticating cookie?
- encrypted HTTP
- unsigned HTTP
- compressed HTTP
- signed HTTP
MapReduce jobs can be written in which language?
- Java or Python
- SQL only
- SQL or Java
- Python or SQL
To perform local aggregation of the intermediate outputs, MapReduce users can optionally specify which object?
- Reducer
- Combiner
- Mapper
- Counter
To verify job status, look for the value ____ in the ____.
- SUCCEEDED; syslog
- SUCCEEDED; stdout
- DONE; syslog
- DONE; stdout
Which line of code implements a Reducer method in MapReduce 2.0?
- public void reduce(Text key, Iterator values, Context context){…}
- public static void reduce(Text key, IntWritable[] values, Context context){…}
- public static void reduce(Text key, Iterator values, Context context){…}
- public void reduce(Text key, IntWritable[] values, Context context){…}
To get the total number of mapped input records in a map job task, you should review the value of which counter?
- FileInputFormatCounter
- FileSystemCounter
- JobCounter
- TaskCounter (NOT SURE)
Hadoop Core supports which CAP capabilities?
- A, P
- C, A
- C, P
- C, A, P
What are the primary phases of a Reducer?
- combine, map, and reduce
- shuffle, sort, and reduce
- reduce, sort, and combine
- map, sort, and combine
To set up Hadoop workflow with synchronization of data between jobs that process tasks both on disk and in memory, use the _____ service, which is _____.
- Oozie; open source
- Oozie; commercial software
- Zookeeper; commercial software
- Zookeeper; open source
For high availability, use multiple nodes of which type?
- data
- name
- memory
- worker
DataNode supports which type of drives?
- hot swappable
- cold swappable
- warm swappable
- non-swappable
Which method is used to implement Spark jobs?
- on disk of all workers
- on disk of the master node
- in memory of the master node
- in memory of all workers
In a MapReduce job, where does the map() function run?
- on the reducer nodes of the cluster
- on the data nodes of the cluster (NOT SURE)
- on the master node of the cluster
- on every node of the cluster
To reference a master file for lookups during Mapping, what type of cache should be used?
- distributed cache
- local cache
- partitioned cache
- cluster cache
Skip bad records provides an option where a certain set of bad input records can be skipped when processing what type of data?
- cache inputs
- reducer inputs
- intermediate values
- map inputs
CREDITS: (Source)
&
Click here to see solutions for all Machine Learning Coursera Assignments.
&
Click here to see more codes for Raspberry Pi 3 and similar Family.
&
Click here to see more codes for NodeMCU ESP8266 and similar Family.
&
Click here to see more codes for Arduino Mega (ATMega 2560) and similar Family.
Feel free to ask doubts in the comment section. I will try my best to answer it.
If you find this helpful by any mean like, comment and share the post.
This is the simplest way to encourage me to keep doing such work.
- APDaga DumpBox