HCAHD Dumps HCAHD Braindumps HCAHD Real Questions HCAHD Practice Test HCAHD Actual Questions Hitachi HCAHD Apache Hadoop Developer https://killexams.com/pass4sure/exam-detail/HCAHD Question: 24 Assuming the following Hive query executes successfully: Which one of the following statements describes the result set? A. A bigram of the top 80 sentences that contain the substring "you are" in the lines column of the input data A1 table. B. An 80-value ngram of sentences that contain the words "you" or "are" in the lines column of the inputdata table. C. A trigram of the top 80 sentences that contain "you are" followed by a null space in the lines column of the inputdata table. D. A frequency distribution of the top 80 words that follow the subsequence "you are" in the lines column of the inputdata table. Answer: D Question: 25 Given the following Pig commands: Which one of the following statements is true? A. The $1 variable represents the first column of data in 'my.log' B. The $1 variable represents the second column of data in 'my.log' C. The severe relation is not valid D. The grouped relation is not valid Answer: B Question: 26 What does Pig provide to the overall Hadoop solution? A. Legacy language Integration with MapReduce framework B. Simple scripting language for writing MapReduce programs C. Database table and storage management services D. C++ interface to MapReduce and data warehouse infrastructure Answer: B Question: 27 What types of algorithms are difficult to express in MapReduce v1 (MRv1)? A. Algorithms that require applying the same mathematical function to large numbers of individual binary records. B. Relational operations on large amounts of structured and semi-structured data. C. Algorithms that require global, sharing states. D. Large-scale graph algorithms that require one-step link traversal. E. Text analysis algorithms on large collections of unstructured text (e.g, Web crawls). Answer: A,C,E Explanation: See 3) below. Limitations of Mapreduce C where not to use Mapreduce While very powerful and applicable to a wide variety of problems, MapReduce is not the answer to every problem. Here are some problems I found where MapReudce is not suited and some papers that address the limitations of MapReuce. Question: 28 You need to create a job that does frequency analysis on input data. You will do this by writing a Mapper that uses TextInputFormat and splits each value (a line of text from an input file) into individual characters. For each one of these characters, you will emit the character as a key and an InputWritable as the value. As this will produce proportionally more intermediate data than input data, which two resources should you expect to be bottlenecks? A. Processor and network I/O B. Disk I/O and network I/O C. Processor and RAM D. Processor and disk I/O Answer: B Question: 29 Which one of the following statements regarding the components of YARN is FALSE? A. A Container executes a specific task as assigned by the ApplicationMaster B. The ResourceManager is responsible for scheduling and allocating resources C. A client application submits a YARW job to the ResourceManager D. The ResourceManager monitors and restarts any failed Containers Answer: D Question: 30 You are developing a combiner that takes as input Text keys, IntWritable values, and emits Text keys, IntWritable values. Which interface should your class implement? A. Combiner B. Mapper C. Reducer D. Reducer E. Combiner Answer: D Question: 31 Which one of the following Hive commands uses an HCatalog table named x? A. SELECT * FROM x; B. SELECT x.-FROM org.apache.hcatalog.hive.HCatLoader('x'); C. SELECT * FROM org.apache.hcatalog.hive.HCatLoader('x'); D. Hive commands cannot reference an HCatalog table Answer: C Question: 32 Given the following Pig command: logevents = LOAD 'input/my.log' AS (date:chararray, levehstring, code:int, message:string); Which one of the following statements is true? A. The logevents relation represents the data from the my.log file, using a comma as the parsing delimiter B. The logevents relation represents the data from the my.log file, using a tab as the parsing delimiter C. The first field of logevents must be a properly-formatted date string or table return an error D. The statement is not a valid Pig command Answer: B Question: 33 Consider the following two relations, A and B. A. C = DOIN B BY a1, A by b2; B. C = JOIN A by al, B by b2; C. C = JOIN A a1, B b2; D. C = JOIN A SO, B $1; Answer: B Question: 34 Given the following Hive commands: Which one of the following statements Is true? A. The file mydata.txt is copied to a subfolder of /apps/hive/warehouse B. The file mydata.txt is moved to a subfolder of /apps/hive/warehouse C. The file mydata.txt is copied into Hive's underlying relational database 0. D. The file mydata.txt does not move from Its current location in HDFS Answer: A Question: 35 In a MapReduce job, the reducer receives all values associated with same key. Which statement best describes the ordering of these values? A. The values are in sorted order. B. The values are arbitrarily ordered, and the ordering may vary from run to run of the same MapReduce job. C. The values are arbitrary ordered, but multiple runs of the same MapReduce job will always have the same ordering. D. Since the values come from mapper outputs, the reducers will receive contiguous sections of sorted values. Answer: A,B Explanation: Note: * Input to the Reducer is the sorted output of the mappers. * The framework calls the application's Reduce function once for each unique key in the sorted order. * Example: For the given sample input the first map emits: < Hello, 1> < World, 1> < Bye, 1> < World, 1> The second map emits: < Hello, 1> < Hadoop, 1> < Goodbye, 1> < Hadoop, 1> Question: 36 Which describes how a client reads a file from HDFS? A. The client queries the NameNode for the block location(s). The NameNode returns the block location(s) to the client. The client reads the data directory off the DataNode(s). B. The client queries all DataNodes in parallel. The DataNode that contains the requested data responds directly to the client. The client reads the data directly off the DataNode. C. The client contacts the NameNode for the block location(s). The NameNode then queries the DataNodes for block locations. The DataNodes respond to the NameNode, and the NameNode redirects the client to the DataNode that holds the requested data block(s). The client then reads the data directly off the DataNode. D. The client contacts the NameNode for the block location(s). The NameNode contacts the DataNode that holds the requested data block. Data is transferred from the DataNode to the NameNode, and then from the NameNode to the client. Answer: A,C,D Explanation: Reference: 24 Interview Questions & Answers for Hadoop MapReduce developers, How the Client communicates with HDFS? Question: 37 For each input key-value pair, mappers can emit: A. As many intermediate key-value pairs as designed. There are no restrictions on the types of those key-value pairs (i.e., they can be heterogeneous). B. As many intermediate key-value pairs as designed, but they cannot be of the same type as the input key-value pair. C. One intermediate key-value pair, of a different type. D. One intermediate key-value pair, but of the same type. E. As many intermediate key-value pairs as designed, as long as all the keys have the same types and all the values have the same type. Answer: A,E Explanation: Mapper maps input key/value pairs to a set of intermediate key/value pairs. Maps are the individual tasks that transform input records into intermediate records. The transformed intermediate records do not need to be of the same type as the input records. A given input pair may map to zero or many output pairs. Reference: Hadoop Map-Reduce Tutorial Question: 38 You write MapReduce job to process 100 files in HDFS. Your MapReduce algorithm uses TextInputFormat: the mapper applies a regular expression over input values and emits key-values pairs with the key consisting of the matching text, and the value containing the filename and byte offset. Determine the difference between setting the number of reduces to one and settings the number of reducers to zero. A. There is no difference in output between the two settings. B. With zero reducers, no reducer runs and the job throws an exception. With one reducer, instances of matching patterns are stored in a single file on HDF C. With zero reducers, all instances of matching patterns are gathered together in one file on HDF D. With one reducer, instances of matching patterns are stored in multiple files on HDF E. With zero reducers, instances of matching patterns are stored in multiple files on HDF F. With one reducer, all instances of matching patterns are gathered together in one file on HDF Answer: A,C,D Explanation: * It is legal to set the number of reduce-tasks to zero if no reduction is desired. In this case the outputs of the map-tasks go directly to the FileSystem, into the output path set by setOutputPath(Path). The framework does not sort the map-outputs before writing them out to the FileSystem. * Often, you may want to process input data using a map function only. To do this, simply set mapreduce.job.reduces to zero. The MapReduce framework will not create any reducer tasks. Rather, the outputs of the mapper tasks will be the final output of the job. Note: Reduce In this phase the reduce(WritableComparable, Iterator, OutputCollector, Reporter) method is called for each pair in the grouped inputs. The output of the reduce task is typically written to the FileSystem via OutputCollector.collect(WritableComparable, Writable). Applications can use the Reporter to report progress, set application-level status messages and update Counters, or just indicate that they are alive. The output of the Reducer is not sorted. Question: 39 In Hadoop 2.0, which one of the following statements is true about a standby NameNode? The Standby NameNode: A. Communicates directly with the active NameNode to maintain the state of the active NameNode. B. Receives the same block reports as the active NameNode. C. Runs on the same machine and shares the memory of the active NameNode. D. Processes all client requests and block reports from the appropriate DataNodes. Answer: B Question: 40 In the reducer, the MapReduce API provides you with an iterator over Writable values. What does calling the next () method return? A. It returns a reference to a different Writable object time. B. It returns a reference to a Writable object from an object pool. C. It returns a reference to the same Writable object each time, but populated with different data. D. It returns a reference to a Writable object. The API leaves unspecified whether this is a reused object or a new object. E. It returns a reference to the same Writable object if the next value is the same as the previous value, or a new Writable object otherwise. Answer: A,C,E Explanation: Calling Iterator.next() will always return the SAME EXACT instance of IntWritable, with the contents of that instance replaced with the next value. Reference: manupulating iterator in mapreduce 6$03/( 48(67,216 7KHVH TXHVWLRQV DUH IRU GHPR SXUSRVH RQO\ )XOO YHUVLRQ LV XS WR GDWH DQG FRQWDLQV DFWXDO TXHVWLRQV DQG DQVZHUV .LOOH[DPV FRP LV DQ RQOLQH SODWIRUP WKDW RIIHUV D ZLGH UDQJH RI VHUYLFHV UHODWHG WR FHUWLILFDWLRQ H[DP SUHSDUDWLRQ 7KH SODWIRUP SURYLGHV DFWXDO TXHVWLRQV H[DP GXPSV DQG SUDFWLFH WHVWV WR KHOS LQGLYLGXDOV SUHSDUH IRU YDULRXV FHUWLILFDWLRQ H[DPV ZLWK FRQILGHQFH +HUH DUH VRPH NH\ IHDWXUHV DQG VHUYLFHV RIIHUHG E\ .LOOH[DPV FRP $FWXDO ([DP 4XHVWLRQV .LOOH[DPV FRP SURYLGHV DFWXDO H[DP TXHVWLRQV WKDW DUH H[SHULHQFHG LQ WHVW FHQWHUV 7KHVH TXHVWLRQV DUH XSGDWHG UHJXODUO\ WR HQVXUH WKH\ DUH XS WR GDWH DQG UHOHYDQW WR WKH ODWHVW H[DP V\OODEXV %\ VWXG\LQJ WKHVH DFWXDO TXHVWLRQV FDQGLGDWHV FDQ IDPLOLDUL]H WKHPVHOYHV ZLWK WKH FRQWHQW DQG IRUPDW RI WKH UHDO H[DP ([DP 'XPSV .LOOH[DPV FRP RIIHUV H[DP GXPSV LQ 3') IRUPDW 7KHVH GXPSV FRQWDLQ D FRPSUHKHQVLYH FROOHFWLRQ RI TXHVWLRQV DQG DQVZHUV WKDW FRYHU WKH H[DP WRSLFV %\ XVLQJ WKHVH GXPSV FDQGLGDWHV FDQ HQKDQFH WKHLU NQRZOHGJH DQG LPSURYH WKHLU FKDQFHV RI VXFFHVV LQ WKH FHUWLILFDWLRQ H[DP 3UDFWLFH 7HVWV .LOOH[DPV FRP SURYLGHV SUDFWLFH WHVWV WKURXJK WKHLU GHVNWRS 9&( H[DP VLPXODWRU DQG RQOLQH WHVW HQJLQH 7KHVH SUDFWLFH WHVWV VLPXODWH WKH UHDO H[DP HQYLURQPHQW DQG KHOS FDQGLGDWHV DVVHVV WKHLU UHDGLQHVV IRU WKH DFWXDO H[DP 7KH SUDFWLFH WHVWV FRYHU D ZLGH UDQJH RI TXHVWLRQV DQG HQDEOH FDQGLGDWHV WR LGHQWLI\ WKHLU VWUHQJWKV DQG ZHDNQHVVHV *XDUDQWHHG 6XFFHVV .LOOH[DPV FRP RIIHUV D VXFFHVV JXDUDQWHH ZLWK WKHLU H[DP GXPSV 7KH\ FODLP WKDW E\ XVLQJ WKHLU PDWHULDOV FDQGLGDWHV ZLOO SDVV WKHLU H[DPV RQ WKH ILUVW DWWHPSW RU WKH\ ZLOO UHIXQG WKH SXUFKDVH SULFH 7KLV JXDUDQWHH SURYLGHV DVVXUDQFH DQG FRQILGHQFH WR LQGLYLGXDOV SUHSDULQJ IRU FHUWLILFDWLRQ H[DPV 8SGDWHG &RQWHQW .LOOH[DPV FRP UHJXODUO\ XSGDWHV LWV TXHVWLRQ EDQN DQG H[DP GXPSV WR HQVXUH WKDW WKH\ DUH FXUUHQW DQG UHIOHFW WKH ODWHVW FKDQJHV LQ WKH H[DP V\OODEXV 7KLV KHOSV FDQGLGDWHV VWD\ XS WR GDWH ZLWK WKH H[DP FRQWHQW DQG LQFUHDVHV WKHLU FKDQFHV RI VXFFHVV 7HFKQLFDO 6XSSRUW .LOOH[DPV FRP SURYLGHV IUHH [ WHFKQLFDO VXSSRUW WR DVVLVW FDQGLGDWHV ZLWK DQ\ TXHULHV RU LVVXHV WKH\ PD\ HQFRXQWHU ZKLOH XVLQJ WKHLU VHUYLFHV 7KHLU FHUWLILHG H[SHUWV DUH DYDLODEOH WR SURYLGH JXLGDQFH DQG KHOS FDQGLGDWHV WKURXJKRXW WKHLU H[DP SUHSDUDWLRQ MRXUQH\ 'PS .PSF FYBNT WJTJU IUUQT LJMMFYBNT DPN WFOEPST FYBN MJTU .LOO \RXU H[DP DW )LUVW $WWHPSW *XDUDQWHHG