[proxy] web.archive.org← back | site home | direct (HTTPS) ↗ | proxy home | ◑ dark◐ light

Delta Lake - Reliable Data Lakes at Scale

The Wayback Machine - https://web.archive.org/web/20220112160736/https://delta.io/

Latest News

WATCH NOW

Delta Sharing

Delta Sharing is the industry’s first open protocol for secure data sharing, making it simple to share data with other organizations regardless of which computing platforms they use.

ACID Transactions

Data lakes typically have multiple data pipelines reading and writing data concurrently, and data engineers have to go through a tedious process to ensure data integrity, due to the lack of transactions. Delta Lake brings ACID transactions to your data lakes. It provides serializability, the strongest level of isolation level. Learn more at Diving into Delta Lake: Unpacking the Transaction Log.

Scalable Metadata Handling

In big data, even the metadata itself can be "big data." Delta Lake treats metadata just like data, leveraging Spark's distributed processing power to handle all its metadata. As a result, Delta Lake can handle petabyte-scale tables with billions of partitions and files at ease.

Time Travel (data versioning)

Open Format

All data in Delta Lake is stored in Apache Parquet format enabling Delta Lake to leverage the efficient compression and encoding schemes that are native to Parquet.

Unified Batch and Streaming Source and Sink

A table in Delta Lake is both a batch table, as well as a streaming source and sink. Streaming data ingest, batch historic backfill, and interactive queries all just work out of the box.

Schema Enforcement

Delta Lake provides the ability to specify your schema and enforce it. This helps ensure that the data types are correct and required columns are present, preventing bad data from causing data corruption. For more information, refer to Diving Into Delta Lake: Schema Enforcement & Evolution.

Schema Evolution

Audit History

Delta Lake transaction log records details about every change made to data providing a full audit trail of the changes.

Updates and Deletes

100% Compatible with Apache Spark API

Developers can use Delta Lake with their existing data pipelines with minimal change as it is fully compatible with Spark, the commonly used big data processing engine.

Delta Everywhere

Use the language, services, connectors, or database of your choice with Delta Lake with connectors including Rust, Python, DBT, Hive, Presto, and more!

Instead of parquet...

dataframe
   .write
   .format("parquet")
   .save("/data")

… simply say delta

dataframe
   .write
   .format("delta")
   .save("/data")

Together, the features of Delta Lake improve both the manageability and performance of working with data in cloud storage objects, and enable a "lakehouse" paradigm that combines the key features of data warehouses and data lakes: standard DBMS management functions usable against low-cost object stores.

Delta Lake is an independent open-source project and not controlled by any single company. To emphasize this we joined the Delta Lake Project in 2019, which is a sub-project of the Linux Foundation Projects.

Within the project, we make decisions based on these rules.