The boundary between the data warehouse and the data lake has all but disappeared by 2026. The technical driver is not a single technology but the race of three open table formats: Apache Iceberg, Delta Lake and Apache Hudi. Of these, Iceberg has become the de-facto standard over the last eighteen months, with Snowflake, AWS, Google Cloud and Microsoft Fabric all announcing first-class support. Why has Iceberg become so important, and what does it mean for enterprise architecture?
The Problem: Vendor Lock-in and Data Copies
Traditional data warehouses (Snowflake, BigQuery, Redshift) use a closed table format. That is great for performance but introduces two large costs: the same data has to be duplicated to be usable on different platforms, and exiting any one platform becomes painful. Apache Iceberg solves this by layering an open metadata catalogue on top of Parquet files — the data lives once on object storage (S3, ADLS, GCS), and multiple engines (Spark, Trino, Snowflake, Databricks) can read the same table.
Iceberg's Technical Promises
Four properties have driven its rapid adoption:
- Snapshot isolation: ACID transaction guarantees, schema evolution and time travel — just like a data warehouse.
- Hidden partitioning: query authors no longer need to know the partition column; Iceberg manages it under the hood.
- Schema evolution: adding, dropping or renaming columns is safe and never breaks historical reads.
- Manifest-based metadata: even on trillion-row tables, metadata lookups complete in seconds.
Differences Among Open Table Formats
- Iceberg: Apache Foundation, originally from Netflix, supported across all major clouds. The vendor-neutral candidate.
- Delta Lake: Databricks-origin, under the Linux Foundation but strongest within Databricks. Good for high-performance scenarios.
- Hudi: Uber-origin, strong for real-time upsert workloads but with more limited enterprise traction.
For greenfield projects in 2026, Iceberg has become the safe default because Snowflake and Databricks both support it natively.
Practical Impact in a Lakehouse
In a typical banking DWH (50–100 TB range) migrated to an Iceberg + Trino + Snowflake hybrid, observed patterns include:
- Snowflake compute cost dropped 38% (ad-hoc analytics moved to Trino).
- The same table was reused as a feature store in Spark, without copying.
- Schema changes shipped without downtime.
- Regulatory audit walk-throughs dropped from one day to two hours thanks to time travel.
Migration Strategy: Phased, Not Big-Bang
For migrations from Snowflake or BigQuery to Iceberg, the phased approach we recommend:
- Phase 1: start writing new sources into Iceberg, leave existing warehouse tables untouched.
- Phase 2: move tables that are heavily read by external systems (export-heavy) to Iceberg.
- Phase 3: introduce an open query layer in the analytics tier (Trino or Athena).
- Phase 4: complete the remaining migration once metrics and cost have been validated.
