This document discusses how Apache Spark can be used in machine learning workflows. It covers typical Spark components and cluster hardware configurations. It then discusses how Spark fits into the ML modeling lifecycle, from loading and preparing data to training, evaluating, and deploying models. Specific examples covered include using Spark for distributed model training, grid search, and batch scoring. The document concludes by summarizing how Spark can be used for large-scale training of single and multiple models and applying models to many inputs.