Product

Improved open-table analytics stack with Iceberg, Polaris, Hudi, Delta Lake

Image showing how we improved our open-table support for Apache Iceberg, Apache Polaris, Apache Hudi, and Delta Lake for faster queries, smoother catalog browsing, ACID, and honoring time travel.

Product Update – Enhanced Open-Table Support

Want to see e6data in action?

Learn how data teams power their workloads.

Get Demo
Get Demo

TL;DR: Connect once, query everywhere. Our latest release provides advanced support for Apache Iceberg, Apache Polaris, Apache Hudi, and Delta Lake, enabling you to run fast SQL queries across every open table format without copying data or rewriting pipelines.

Why does this matter for your lakehouse?

Open table formats have become the backbone of modern lakehouses. Iceberg delivers petabyte-scale performance and versioned metadata. Polaris adds a cloud-native catalog with enterprise-grade governance, while Delta Lake brings ACID guarantees and time-travel to the data lake. Teams want the freedom to mix these technologies without the pain of managing multiple engines, drivers, and security models.

With this improved support, e6data’s query engine supports all four dialects natively, consolidating your existing catalogs into a single workspace.

What’s new in this release?

Format What e6data now does Biggest win
Apache Polaris Create, edit, and delete Polaris catalog connections in the UI, with full Iceberg table exploration (schema discovery, partitioned queries) and improved metadata sync + catalog validation during setup. Centralized governance with cloud-native scale.
Apache Iceberg Reads any Iceberg table exposed through Snowflake Open Catalog, AWS Glue, or Polaris with support for partitioned and non‑partitioned tables, enhanced schema evolution and metadata‑aware querying, plus an improved catalog management UI for editing and managing integrations. Query massive, partitioned Iceberg datasets in seconds—no extra Spark job needed.
Apache Hudi Reads any Hudi table via AWS Glue or Apache Hive catalogs, with CoW/MoR support, read‑optimized / incremental / real‑time (MoR) query modes, time travel, schema evolution, upserts/inserts/deletes, and access to commit metadata for incremental processing. External metastore integration is improved for smoother table discovery. CDC‑style incremental queries in seconds using commit metadata — ideal for upsert‑heavy MoR workloads.
Delta Lake Reads any Delta Lake table via AWS Glue or Unity Catalog, with improved schema/file‑stats visibility, a cleaner catalog selection & validation UI, and a browsing/querying experience aligned with other table formats. Keep your Delta pipelines as-is while unlocking sub-second SQL in e6data.

How it works (architecture and performance path)

  • Smart Catalog Layer – e6data now understands Iceberg and Delta layout metadata natively and communicates with Polaris through the same REST interface Iceberg uses.
  • Zero-copy ingestion – Data stays in your object-storage (be it S3, GCS, or ADLS). We only read the files your query touches, reducing scan costs and time.
  • Cross-catalog joins – Need to join a Delta customer table with an Iceberg events log? One SQL query across all.

Fine-grained security – We respect Polaris RBAC and Unity Catalog privileges automatically, so users see only what they’re allowed to see.

Quick start

  1. Create or select a workspace in the e6data console.
  2. Add catalog → Polaris / Iceberg / Hudi / Delta (choose one or all)
  3. Run SHOW TABLES FROM <catalog>.<namespace>; to confirm visibility.
  4. Point your BI tool (Tableau, Power BI, Looker) at the e6data endpoint and start exploring.

Reference links

  1. Apache Polaris documentation
  2. Apache Iceberg documentation
  3. Apache Hudi documentation
  4. Delta Lake documentation
     

Roadmap: write support and deeper governance

  • Write-support for Iceberg and Delta—including MERGE, UPDATE, and DELETE—is on the roadmap.
  • Table-level RBAC in Polaris and automatic metadata sync across engines are in active development.

Stay in the know - Data Engineering ACID

3 Curated stories each week on data engineering at scale—handpicked by the e6data team.

Subscribe Now
Share on

Build future-proof data products

Try e6data for your heavy workloads!

Get Started for Free
Get Started for Free
Frequently asked questions (FAQs)
How do I integrate e6data with my existing data infrastructure?

We are universally interoperable and open-source friendly. We can integrate across any object store, table format, data catalog, governance tools, BI tools, and other data applications.

How does billing work?

We use a usage-based pricing model based on vCPU consumption. Your billing is determined by the number of vCPUs used, ensuring you only pay for the compute power you actually consume.

What kind of file formats does e6data support?

We support all types of file formats, like Parquet, ORC, JSON, CSV, AVRO, and others.

What kind of performance improvements can I expect with e6data?

e6data promises a 5 to 10 times faster querying speed across any concurrency at over 50% lower total cost of ownership across the workloads as compared to any compute engine in the market.

What kinds of deployment models are available at e6data ?

We support serverless and in-VPC deployment models. 

How does e6data handle data governance rules?

We can integrate with your existing governance tool, and also have an in-house offering for data governance, access control, and security.

Table of contents:
Listen to the full podcast
Apple Podcasts
Spotify

Subscribe to our newsletter - Data Engineering ACID

Get 3 weekly stories around data engineering at scale that the e6data team is reading.
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
Share this article

Improved open-table analytics stack with Iceberg, Polaris, Hudi, Delta Lake

July 24, 2025
/
e6data Team
Product
Polaris, Hudi, Delta Lake — read Iceberg/Hudi/Delta via Open/Glue/Polaris; browse catalogs; ACID + time travel (MoR)

TL;DR: Connect once, query everywhere. Our latest release provides advanced support for Apache Iceberg, Apache Polaris, Apache Hudi, and Delta Lake, enabling you to run fast SQL queries across every open table format without copying data or rewriting pipelines.

Why does this matter for your lakehouse?

Open table formats have become the backbone of modern lakehouses. Iceberg delivers petabyte-scale performance and versioned metadata. Polaris adds a cloud-native catalog with enterprise-grade governance, while Delta Lake brings ACID guarantees and time-travel to the data lake. Teams want the freedom to mix these technologies without the pain of managing multiple engines, drivers, and security models.

With this improved support, e6data’s query engine supports all four dialects natively, consolidating your existing catalogs into a single workspace.

What’s new in this release?

Format What e6data now does Biggest win
Apache Polaris Create, edit, and delete Polaris catalog connections in the UI, with full Iceberg table exploration (schema discovery, partitioned queries) and improved metadata sync + catalog validation during setup. Centralized governance with cloud-native scale.
Apache Iceberg Reads any Iceberg table exposed through Snowflake Open Catalog, AWS Glue, or Polaris with support for partitioned and non‑partitioned tables, enhanced schema evolution and metadata‑aware querying, plus an improved catalog management UI for editing and managing integrations. Query massive, partitioned Iceberg datasets in seconds—no extra Spark job needed.
Apache Hudi Reads any Hudi table via AWS Glue or Apache Hive catalogs, with CoW/MoR support, read‑optimized / incremental / real‑time (MoR) query modes, time travel, schema evolution, upserts/inserts/deletes, and access to commit metadata for incremental processing. External metastore integration is improved for smoother table discovery. CDC‑style incremental queries in seconds using commit metadata — ideal for upsert‑heavy MoR workloads.
Delta Lake Reads any Delta Lake table via AWS Glue or Unity Catalog, with improved schema/file‑stats visibility, a cleaner catalog selection & validation UI, and a browsing/querying experience aligned with other table formats. Keep your Delta pipelines as-is while unlocking sub-second SQL in e6data.

How it works (architecture and performance path)

  • Smart Catalog Layer – e6data now understands Iceberg and Delta layout metadata natively and communicates with Polaris through the same REST interface Iceberg uses.
  • Zero-copy ingestion – Data stays in your object-storage (be it S3, GCS, or ADLS). We only read the files your query touches, reducing scan costs and time.
  • Cross-catalog joins – Need to join a Delta customer table with an Iceberg events log? One SQL query across all.

Fine-grained security – We respect Polaris RBAC and Unity Catalog privileges automatically, so users see only what they’re allowed to see.

Quick start

  1. Create or select a workspace in the e6data console.
  2. Add catalog → Polaris / Iceberg / Hudi / Delta (choose one or all)
  3. Run SHOW TABLES FROM <catalog>.<namespace>; to confirm visibility.
  4. Point your BI tool (Tableau, Power BI, Looker) at the e6data endpoint and start exploring.

Reference links

  1. Apache Polaris documentation
  2. Apache Iceberg documentation
  3. Apache Hudi documentation
  4. Delta Lake documentation
     

Roadmap: write support and deeper governance

  • Write-support for Iceberg and Delta—including MERGE, UPDATE, and DELETE—is on the roadmap.
  • Table-level RBAC in Polaris and automatic metadata sync across engines are in active development.

Stay in the know - Data Engineering ACID

3 Curated stories each week on data engineering at scale—handpicked by the e6data team.

Subscribe Now
Listen to the full podcast
Share this article

FAQs

How does e6data reduce Snowflake compute costs without slowing queries?
e6data is powered by the industry’s only atomic architecture. Rather than scaling in step jumps (L x 1 -> L x 2), e6data scales atomically, by as little as 1 vCPU. In production with widely varying loads, this translates to > 60% TCO savings.
Do I have to move out of Snowflake?
No, we fit right into your existing data architecture across cloud, on-prem, catalog, governance, table formats, BI tools, and more.

Does e6data speed up Iceberg on Snowflake?
Yes, depending on your workload, you can see anywhere up to 10x faster speeds through our native and advanced Iceberg support. 

Snowflake supports Iceberg. But how do you get data there in real time?
Our real-time streaming ingest streams Kafka or SDK data straight into Iceberg—no Flink. Landing within 60 seconds and auto-registering each snapshot for instant querying.

How long does it take to deploy e6data alongside Snowflake?
Sign up the form and get your instance started. You can deploy it to any cloud, region, deployment model, without copying or migrating any data from Snowflake.

FAQs

What open table formats does the new e6data release support?
The July 2025 update adds native, side-by-side support for Apache Iceberg, Apache Polaris, Apache Hudi, and Delta Lake, letting you query any table written in those four formats through a single e6data workspace.
Why is a consolidated catalog useful for my lakehouse?
By hooking Iceberg, Polaris, Hudi, and Delta catalogs into one workspace, you avoid juggling multiple drivers, security models, or UIs while gaining consistent schema discovery, partition browsing, and cross-catalog joins.
Are my existing governance rules respected?
Yes, e6data automatically enforces Polaris RBAC and Unity Catalog privileges, ensuring users only see the tables and columns they’re entitled to query.
Will I need to change existing Delta or Iceberg pipelines?
No, the update lets you keep your current pipelines “as-is”; e6data simply reads the tables and serves faster SQL without extra Spark jobs.

Related posts

View All Posts

Related posts

View All
Engineering
This is some text inside of a div block.
August 8, 2025
/
Adishesh Kishore
Embedding Essentials: From Cosine Similarity to SQL with Vectors
Adishesh Kishore
August 8, 2025
View All
Product
This is some text inside of a div block.
August 5, 2025
/
e6data Team
e6data’s Hybrid Data Lakehouse: 10x Faster Queries, Near-Zero Egress, Sub-Second Latency
e6data Team
August 5, 2025
View All
Engineering
This is some text inside of a div block.
July 25, 2025
/
Rajath Gowda
Building a Modern Data Pipeline in Snowflake: From Snowpipe to Managed Iceberg Tables with Sync Checks
Rajath Gowda
July 25, 2025
View All Posts