Distributed Storage and Processing Framework
Framework that allows for distributed processing of large datasets across clusters of computers using simple programming models. Designed to scale from single servers to thousands of machines, each offering local computation and storage. Uses HDFS for distributed storage and MapReduce for processing.
Unified Batch and Stream Processing
Advanced unified programming model for defining and executing data processing workflows that can run on any execution engine. Provides portability across multiple execution environments including Apache Flink, Apache Spark, and Google Cloud Dataflow. Ideal for building flexible, scalable data pipelines.