Delta Lake is an open storage layer that brings reliability to data lakes. By providing ACID transactions and data versioning, Delta Lake enables you to perform multiple reads, writes, and merges on tables with full confidence. Delta Lake is the first unified storage system capable of ingesting, consolidating, and managing structured and semi-structured (JSON) data with high performance for all your real-time machine learning and analytics use cases. Delta Lake helps save time by resolving issues such as duplicated data, incomplete updates, and corruption before they impact downstream applications. Delta Lake enables you to have a single source of truth for all of your enterprise data with its unified platform for batch and streaming workloads. Delta Lake integrates streaming & batch data processing while providing ACID transactions and scalable metadata handling.
The Delta Lake project is now an open-source project in the Linux Foundation and is the basis for a set of upcoming features in Databricks Runtime. Since Delta Lake was launched in October 2017, Delta Lake has been adopted by over 4,000 organizations and processes over two exabytes of data each month.