Apache Spark is all the rage these days. People who work with Big Data, Spark is a household name for them. We have been using it for quite some time now. So we already know that Spark is lightning-fast cluster computing technology, it is faster than Hadoop MapReduce. If you ask any of these Spark techies, how Spark is fast, they would give you a vague answer by saying Spark uses DAG to carry out the in-memory computations. So, how far is this answer satisfiable? Well to a Spark expert, this answer is just equivalent to a poison. Let’s try to understand how exactly spark is handling our computations through DAG.