Which Apache project is used for workflow automation in Hadoop?
A. Oozie
B. Flume
C. Spark
D. NiFi
Answer: Option A
A. Oozie
B. Flume
C. Spark
D. NiFi
Answer: Option A
What is Hadoop primarily designed for?
A. Real-time data processing
B. Batch processing of large datasets
C. Structured data storage
D. In-memory caching
What is the core component of Hadoop responsible for distributed storage?
A. YARN
B. HDFS
C. MapReduce
D. Hive
Which programming language is commonly used for writing MapReduce programs?
A. Python
B. Java
C. Ruby
D. C++
What does HDFS stand for in the context of Hadoop?
A. Hadoop Distributed File System
B. High-level Data Format System
C. Hadoop Data Flow System
D. High-Density File Storage
Join The Discussion