This document discusses using distributed systems like Hadoop for processing large data sets known as "Big Data". It introduces key concepts like MapReduce, HDFS, HBase and Pig which are used to build distributed systems on commodity hardware and software. The presentation covers motivations for using these systems, outlines their basic architectures and limitations, and provides examples and external resources for further learning.