Apache Hadoop is an open source framework that is used to efficiently store and process large datasets ranging in size from gigabytes to petabytes of data. Instead of using one large computer to store and process the data, Hadoop allows clustering multiple computers to analyze massive datasets in parallel more quickly.
Get Big Data and Hadoop certificate from The Digital Adda which you can share in the Certifications section of your LinkedIn profile, on printed resumes, CVs, or other documents.
TERMS AND CONDITIONS:
- There are 20 MCQ Questions in this test
- You need to score atleast 50% to get a certificate
- A certificate of achievement will be awarded .
Question 1: Which type of data Hadoop can deal with is
- Structured
- Semi – structured
- Unstructured
- All of the above
Question 2: Hadoop Framework is written in
- Python
- Java
- C++
- Scala
Question 3: Which of the following platforms does Apache Hadoop run on?
- Bare metal
- Unix-like
- Cross-platform
- Debian
Question 4: Which of the following is true about Hadoop?
- Hadoop MapReduce processes the data at very low latency
- Hadoop can’t handle structured data
- Hadoop is schema-less – we don’t need to define a schema before writing the data
- None of the above
Question 5: Which of the following is the correct statement:
- Data locality means moving computation to data instead of data to computation
- Data locality means moving data to computation instead of computation to data
- Both the above
- None of the above
Question 6: Which one of the following is false about Hadoop?
- It is a distributed framework
- The main algorithm used in Hadoop is Map Reduce
- Hadoop can work with commodity hardware
- All are true
Question 7: Which of the following is component of Hadoop?
- YARN
- HDFS
- MapReduce
- All of the above
Question 8: What license is Apache Hadoop distributed under?
- Apache License 2.0
- Shareware
- Mozilla Public License
- Commercial
Question 9: Which of the following is the daemon of Hadoop?
- NameNode
- Node manager
- DataNode
- All of the above
Question 10: Apache Hadoop achieves reliability by replicating the data across multiple hosts, and hence does not require __ storage on hosts.
- Filesystem
- RAID
- Local FS
- Operating system
Question 11: Hadoop works in
- Centralized processing fashion
- master – slave fashion
- worker/slave fashion
- All of the mentioned
Question 12: Which of the following property gets configured on mapred-site.xml?
- Replication factor
- Java Environment variables
- Directory names to store hdfs files
- Which framework to use with MapReduce jobs
Question 13: The archive file created in Hadoop has the extension of
- .hrh
- .har
- .hrc
- .hrar
Question 14: Which of the below component deals with ingesting streaming data into Hadoop?
- Flume
- Oozie
- Hive
- Kafka
Question 15: As compared to RDBMS, Apache Hadoop
- Has higher data Integrity
- Does ACID transactions
- Is suitable for read and write many times
- Works better on unstructured and semi-structured data
Question 16: In which all languages you can code in Hadoop?
- Java
- Python
- C++
- All of the above
Question 17: Hadoop command to list all the blocks corresponding to each file in the hdfs?
- hadoop fsck / -files -blocks
- hadoop fsck / -blocks -files
- hadoop ls / -blocks -files
- hadoop fchk / -files -blocks
Question 18: Zookeeper is
- Synchronization & coordination engine
- NoSQL database
- Data collection & aggregation system
- Data processing engine
Question 19: All of the following accurately describe Hadoop, EXCEPT
- Open source
- Real-time data processing system
- Java-based
- Distributed computing approach
Question 20: Which tool is used to move data from RDBMS to HDFS?
- Sqoop
- Flume
- Oozie
- Drill