"The core idea of Spark is to create a large in-memory data set across a cluster of computers. If HDFS strove to create a single persistent filesystem across a cluster, Spark effectively creates one large memory space across the cluster. At the core of Spark is the Resilient Distributed Dataset (RDD), which appears as a single data set to the programs using Spark."
via Check out this quote from The Enterprise Big Data Lake - https://learning.oreilly.com/library/view/-/9781491931547/ch03.html
SparkとRedisの違いはなんだろうな。