Hadoop vs Apache Spark
The Apache Hadoop software library is a framework that allows distributed processing of large datasets across clusters of computers using simple programming models. Hadoop can be easily
scaled-up to multi cluster machines, each offering local storage and computation. Hadoop libraries are designed in such a way that it can detect the failed cluster at application layer and can handle those failures by it. This ensures high-availability by default.