The “processing frameworks” are one of the most essential components of a Big Data systems. There are three categories of such frameworks namely: Batch-only frameworks (Hadoop), Stream-only frameworks (Storm, Samza), and Hybrid frameworks (Spark, Hive and Flink). In this lecture, we will introduce them and cover one of the major Big Data frameworks, Apache Spark. We will cover Spark fundamentals and the model of “Resilient Distributed Datasets (RDDs)” that are used in Spark to implement in-memory batch computation.
This lecture focuses on architecting Big Data solution. We will discuss the role and importance of the components in realizing system architectures. The participants will be introduced to unique problem characteristics that drive Big Data and the unending technology options to solve them. The application of the introduced concepts and components will be discussed in real-world example of practical use-cases.
This lecture will cover the existing advanced Big Data architectures following a bottom-up approach. In this lecture, the important knowledge to design and architect scalable solutions for challenging problems will be introduced. The primary components in the architecture of such systems and their architectures will be presented and discussed including “inter alia distributed kernels” and cluster managers, distributed file systems and storage systems.