Labour Day Special - Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: top65certs

Hortonworks HDPCD Dumps

Page: 1 / 4
Total 108 questions

Hortonworks Data Platform Certified Developer Questions and Answers

Question 1

Consider the following two relations, A and B.

Which Pig statement combines A by its first field and B by its second field?

Options:

A.

C = DOIN B BY a1, A by b2;

B.

C = JOIN A by al, B by b2;

C.

C = JOIN A a1, B b2;

D.

C = JOIN A SO, B $1;

Question 2

Review the following data and Pig code:

What command to define B would produce the output (M,62,95l02) when invoking the DUMP operator on B?

Options:

A.

B = FILTER A BY (zip = = '95102' AND gender = = M");

B.

B= FOREACH A BY (gender = = 'M' AND zip = = '95102');

C.

B = JOIN A BY (gender = = 'M' AND zip = = '95102');

D.

B= GROUP A BY (zip = = '95102' AND gender = = 'M');

Question 3

You have just executed a MapReduce job. Where is intermediate data written to after being emitted from the Mapper’s map method?

Options:

A.

Intermediate data in streamed across the network from Mapper to the Reduce and is never written to disk.

B.

Into in-memory buffers on the TaskTracker node running the Mapper that spill over and are written into HDFS.

C.

Into in-memory buffers that spill over to the local file system of the TaskTracker node running the Mapper.

D.

Into in-memory buffers that spill over to the local file system (outside HDFS) of the TaskTracker node running the Reducer

E.

Into in-memory buffers on the TaskTracker node running the Reducer that spill over and are written into HDFS.

Question 4

A NameNode in Hadoop 2.2 manages ______________.

Options:

A.

Two namespaces: an active namespace and a backup namespace

B.

A single namespace

C.

An arbitrary number of namespaces

D.

No namespaces

Question 5

Can you use MapReduce to perform a relational join on two large tables sharing a key? Assume that the two tables are formatted as comma-separated files in HDFS.

Options:

A.

Yes.

B.

Yes, but only if one of the tables fits into memory

C.

Yes, so long as both tables fit into memory.

D.

No, MapReduce cannot perform relational operations.

E.

No, but it can be done with either Pig or Hive.

Question 6

MapReduce v2 (MRv2/YARN) splits which major functions of the JobTracker into separate daemons? Select two.

Options:

A.

Heath states checks (heartbeats)

B.

Resource management

C.

Job scheduling/monitoring

D.

Job coordination between the ResourceManager and NodeManager

E.

Launching tasks

F.

Managing file system metadata

G.

MapReduce metric reporting

Question 7

What is the disadvantage of using multiple reducers with the default HashPartitioner and distributing your workload across you cluster?

Options:

A.

You will not be able to compress the intermediate data.

B.

You will longer be able to take advantage of a Combiner.

C.

By using multiple reducers with the default HashPartitioner, output files may not be in globally sorted order.

D.

There are no concerns with this approach. It is always advisable to use multiple reduces.

Question 8

Workflows expressed in Oozie can contain:

Options:

A.

Sequences of MapReduce and Pig. These sequences can be combined with other actions including forks, decision points, and path joins.

B.

Sequences of MapReduce job only; on Pig on Hive tasks or jobs. These MapReduce sequences can be combined with forks and path joins.

C.

Sequences of MapReduce and Pig jobs. These are limited to linear sequences of actions with exception handlers but no forks.

D.

Iterntive repetition of MapReduce jobs until a desired answer or state is reached.

Question 9

Which one of the following files is required in every Oozie Workflow application?

Options:

A.

job.properties

B.

Config-default.xml

C.

Workflow.xml

D.

Oozie.xml

Question 10

What types of algorithms are difficult to express in MapReduce v1 (MRv1)?

Options:

A.

Algorithms that require applying the same mathematical function to large numbers of individual binary records.

B.

Relational operations on large amounts of structured and semi-structured data.

C.

Algorithms that require global, sharing states.

D.

Large-scale graph algorithms that require one-step link traversal.

E.

Text analysis algorithms on large collections of unstructured text (e.g, Web crawls).

Question 11

When can a reduce class also serve as a combiner without affecting the output of a MapReduce program?

Options:

A.

When the types of the reduce operation’s input key and input value match the types of the reducer’s output key and output value and when the reduce operation is both communicative and associative.

B.

When the signature of the reduce method matches the signature of the combine method.

C.

Always. Code can be reused in Java since it is a polymorphic object-oriented programming language.

D.

Always. The point of a combiner is to serve as a mini-reducer directly after the map phase to increase performance.

E.

Never. Combiners and reducers must be implemented separately because they serve different purposes.

Question 12

Which one of the following statements is FALSE regarding the communication between DataNodes and a federation of NameNodes in Hadoop 2.2?

Options:

A.

Each DataNode receives commands from one designated master NameNode.

B.

DataNodes send periodic heartbeats to all the NameNodes.

C.

Each DataNode registers with all the NameNodes.

D.

DataNodes send periodic block reports to all the NameNodes.

Question 13

Which project gives you a distributed, Scalable, data store that allows you random, realtime read/write access to hundreds of terabytes of data?

Options:

A.

HBase

B.

Hue

C.

Pig

D.

Hive

E.

Oozie

F.

Flume

G.

Sqoop

Question 14

To process input key-value pairs, your mapper needs to lead a 512 MB data file in memory. What is the best way to accomplish this?

Options:

A.

Serialize the data file, insert in it the JobConf object, and read the data into memory in the configure method of the mapper.

B.

Place the data file in the DistributedCache and read the data into memory in the map method of the mapper.

C.

Place the data file in the DataCache and read the data into memory in the configure method of the mapper.

D.

Place the data file in the DistributedCache and read the data into memory in the configure method of the mapper.

Question 15

For each intermediate key, each reducer task can emit:

Options:

A.

As many final key-value pairs as desired. There are no restrictions on the types of those key-value pairs (i.e., they can be heterogeneous).

B.

As many final key-value pairs as desired, but they must have the same type as the intermediate key-value pairs.

C.

As many final key-value pairs as desired, as long as all the keys have the same type and all the values have the same type.

D.

One final key-value pair per value associated with the key; no restrictions on the type.

E.

One final key-value pair per key; no restrictions on the type.

Question 16

Which of the following tool was designed to import data from a relational database into HDFS?

Options:

A.

HCatalog

B.

Sqoop

C.

Flume

D.

Ambari

Page: 1 / 4
Total 108 questions