This document defines 16 basic terms related to data engineering: 1. Apache Airflow is an open-source workflow management platform that uses directed acyclic graphs to manage workflow orchestration. 2. Batch processing involves processing large amounts of data at once, such as in ETL steps or bulk operations on digital images. 3. Cold data storage stores old, hardly used data on low-power servers, making retrieval slower. 4. A cluster groups several computers together to perform a single task.