Data Lakehouse
A data lakehouse is a data storage architecture that combines the flexibility and low cost of a data lake (raw files in an S3-style object store) with the transactional and analytical capabilities of a data warehouse (A…
A [data](/ressources/glossaire-de-la-tech/data-donnee) lakehouse is a data storage architecture that combines the flexibility and low cost of a data lake (raw files in an S3-style object store) with the transactional and analytical capabilities of a data warehouse (ACID, schemas, performant SQL queries).
The lakehouse relies on open table formats such as Apache Iceberg, Delta Lake or Apache Hudi, which add a metadata layer on top of Parquet files to offer transactions, time travel and schema evolution.
Reference platforms include Databricks (Delta Lake), Snowflake (Iceberg), AWS Athena and Iceberg-native deployments on Trino, DuckDB or ClickHouse. The lakehouse has become, in 2026, the dominant pattern for analytics at scale.
